var/home/core/zuul-output/0000755000175000017500000000000015114740232014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114752322015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004731775415114752314017722 0ustar rootrootDec 06 05:39:19 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 05:39:19 crc restorecon[4708]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:39:19 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 05:39:20 crc restorecon[4708]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 05:39:20 crc kubenswrapper[4957]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:39:20 crc kubenswrapper[4957]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 05:39:20 crc kubenswrapper[4957]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:39:20 crc kubenswrapper[4957]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:39:20 crc kubenswrapper[4957]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 05:39:20 crc kubenswrapper[4957]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.485801 4957 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489257 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489276 4957 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489281 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489286 4957 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489290 4957 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489318 4957 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489324 4957 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489341 4957 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489348 4957 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489353 4957 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489359 4957 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489364 4957 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489370 4957 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489375 4957 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489380 4957 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489386 4957 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489390 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489394 4957 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489399 4957 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489404 4957 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489408 4957 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489413 4957 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489417 4957 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489421 4957 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489425 4957 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489430 4957 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489433 4957 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489437 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489442 4957 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489447 4957 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489451 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489456 4957 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489465 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489469 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489473 4957 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489486 4957 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489491 4957 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489496 4957 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489500 4957 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489505 4957 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489510 4957 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489514 4957 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489518 4957 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489523 4957 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489528 4957 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489534 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489540 4957 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489544 4957 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489549 4957 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489554 4957 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489559 4957 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489565 4957 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489569 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489572 4957 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489576 4957 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489580 4957 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489584 4957 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489587 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489591 4957 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489595 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489601 4957 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489605 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489608 4957 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489613 4957 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489619 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489622 4957 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489626 4957 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489630 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489633 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489637 4957 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.489641 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489716 4957 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489725 4957 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489733 4957 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489739 4957 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489745 4957 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489750 4957 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489756 4957 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489762 4957 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489767 4957 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489772 4957 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489777 4957 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489781 4957 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489785 4957 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489790 4957 flags.go:64] FLAG: --cgroup-root="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489795 4957 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489800 4957 flags.go:64] FLAG: --client-ca-file="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489804 4957 flags.go:64] FLAG: --cloud-config="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489809 4957 flags.go:64] FLAG: --cloud-provider="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489813 4957 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489818 4957 flags.go:64] FLAG: --cluster-domain="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489822 4957 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489843 4957 flags.go:64] FLAG: --config-dir="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489848 4957 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489853 4957 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489859 4957 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489863 4957 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489867 4957 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489872 4957 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489876 4957 flags.go:64] FLAG: --contention-profiling="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489880 4957 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489885 4957 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489889 4957 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489893 4957 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489899 4957 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489904 4957 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489908 4957 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489913 4957 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489917 4957 flags.go:64] FLAG: --enable-server="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489921 4957 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489929 4957 flags.go:64] FLAG: --event-burst="100" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489934 4957 flags.go:64] FLAG: --event-qps="50" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489938 4957 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489943 4957 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489947 4957 flags.go:64] FLAG: --eviction-hard="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489953 4957 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489957 4957 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489961 4957 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489966 4957 flags.go:64] FLAG: --eviction-soft="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489970 4957 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489974 4957 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489978 4957 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489983 4957 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489987 4957 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489991 4957 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.489995 4957 flags.go:64] FLAG: --feature-gates="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490000 4957 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490004 4957 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490008 4957 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490013 4957 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490017 4957 flags.go:64] FLAG: --healthz-port="10248" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490021 4957 flags.go:64] FLAG: --help="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490025 4957 flags.go:64] FLAG: --hostname-override="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490029 4957 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490033 4957 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490037 4957 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490043 4957 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490047 4957 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490051 4957 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490055 4957 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490059 4957 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490063 4957 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490068 4957 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490072 4957 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490076 4957 flags.go:64] FLAG: --kube-reserved="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490080 4957 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490084 4957 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490088 4957 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490092 4957 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490096 4957 flags.go:64] FLAG: --lock-file="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490101 4957 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490105 4957 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490109 4957 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490115 4957 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490119 4957 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490123 4957 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490127 4957 flags.go:64] FLAG: --logging-format="text" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490131 4957 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490135 4957 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490139 4957 flags.go:64] FLAG: --manifest-url="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490143 4957 flags.go:64] FLAG: --manifest-url-header="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490149 4957 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490153 4957 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490158 4957 flags.go:64] FLAG: --max-pods="110" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490162 4957 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490167 4957 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490171 4957 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490175 4957 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490179 4957 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490184 4957 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490188 4957 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490199 4957 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490203 4957 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490208 4957 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490213 4957 flags.go:64] FLAG: --pod-cidr="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490217 4957 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490224 4957 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490228 4957 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490232 4957 flags.go:64] FLAG: --pods-per-core="0" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490236 4957 flags.go:64] FLAG: --port="10250" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490240 4957 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490244 4957 flags.go:64] FLAG: --provider-id="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490248 4957 flags.go:64] FLAG: --qos-reserved="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490252 4957 flags.go:64] FLAG: --read-only-port="10255" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490256 4957 flags.go:64] FLAG: --register-node="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490260 4957 flags.go:64] FLAG: --register-schedulable="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490265 4957 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490272 4957 flags.go:64] FLAG: --registry-burst="10" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490276 4957 flags.go:64] FLAG: --registry-qps="5" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490280 4957 flags.go:64] FLAG: --reserved-cpus="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490284 4957 flags.go:64] FLAG: --reserved-memory="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490293 4957 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490298 4957 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490303 4957 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490309 4957 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490313 4957 flags.go:64] FLAG: --runonce="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490318 4957 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490322 4957 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490327 4957 flags.go:64] FLAG: --seccomp-default="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490332 4957 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490336 4957 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490341 4957 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490345 4957 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490349 4957 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490353 4957 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490357 4957 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490361 4957 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490365 4957 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490369 4957 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490373 4957 flags.go:64] FLAG: --system-cgroups="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490377 4957 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490383 4957 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490387 4957 flags.go:64] FLAG: --tls-cert-file="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490391 4957 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490396 4957 flags.go:64] FLAG: --tls-min-version="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490400 4957 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490404 4957 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490408 4957 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490412 4957 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490416 4957 flags.go:64] FLAG: --v="2" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490421 4957 flags.go:64] FLAG: --version="false" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490427 4957 flags.go:64] FLAG: --vmodule="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490434 4957 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490440 4957 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490533 4957 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490538 4957 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490543 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490546 4957 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490550 4957 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490554 4957 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490557 4957 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490561 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490565 4957 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490569 4957 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490572 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490577 4957 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490581 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490585 4957 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490589 4957 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490592 4957 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490596 4957 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490600 4957 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490603 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490607 4957 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490610 4957 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490614 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490617 4957 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490621 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490624 4957 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490629 4957 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490633 4957 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490637 4957 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490641 4957 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490645 4957 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490648 4957 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490654 4957 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490657 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490661 4957 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490665 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490669 4957 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490673 4957 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490676 4957 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490680 4957 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490683 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490687 4957 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490695 4957 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490699 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490702 4957 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490706 4957 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490709 4957 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490713 4957 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490717 4957 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490720 4957 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490724 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490727 4957 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490732 4957 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490736 4957 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490739 4957 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490743 4957 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490747 4957 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490750 4957 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490753 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490758 4957 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490763 4957 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490767 4957 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490771 4957 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490774 4957 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490780 4957 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490783 4957 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490787 4957 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490790 4957 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490794 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490797 4957 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490801 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.490806 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.490819 4957 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.500737 4957 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.500770 4957 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500912 4957 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500924 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500931 4957 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500938 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500944 4957 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500949 4957 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500955 4957 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500960 4957 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500968 4957 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500977 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500984 4957 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500989 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.500995 4957 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501001 4957 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501006 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501012 4957 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501017 4957 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501022 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501027 4957 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501032 4957 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501038 4957 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501043 4957 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501049 4957 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501054 4957 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501059 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501065 4957 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501070 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501076 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501081 4957 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501087 4957 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501092 4957 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501099 4957 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501104 4957 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501111 4957 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501119 4957 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501125 4957 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501130 4957 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501136 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501141 4957 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501146 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501152 4957 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501159 4957 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501166 4957 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501172 4957 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501178 4957 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501183 4957 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501189 4957 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501194 4957 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501199 4957 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501204 4957 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501210 4957 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501215 4957 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501220 4957 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501225 4957 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501231 4957 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501236 4957 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501241 4957 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501249 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501255 4957 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501261 4957 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501266 4957 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501271 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501318 4957 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501326 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501332 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501337 4957 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501344 4957 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501350 4957 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501355 4957 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501361 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501367 4957 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.501377 4957 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501558 4957 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501569 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501576 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501583 4957 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501590 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501596 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501602 4957 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501609 4957 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501616 4957 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501624 4957 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501630 4957 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501636 4957 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501641 4957 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501647 4957 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501654 4957 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501659 4957 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501665 4957 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501672 4957 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501679 4957 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501685 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501691 4957 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501697 4957 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501703 4957 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501711 4957 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501718 4957 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501725 4957 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501730 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501736 4957 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501741 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501747 4957 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501753 4957 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501758 4957 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501763 4957 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501768 4957 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501774 4957 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501781 4957 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501786 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501791 4957 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501796 4957 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501801 4957 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501807 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501812 4957 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501817 4957 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501823 4957 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501828 4957 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501857 4957 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501862 4957 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501867 4957 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501873 4957 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501878 4957 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501884 4957 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501889 4957 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501894 4957 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501900 4957 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501905 4957 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501910 4957 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501916 4957 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501923 4957 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501929 4957 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501937 4957 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501942 4957 feature_gate.go:330] unrecognized feature gate: Example Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501948 4957 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501953 4957 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501959 4957 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501964 4957 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501969 4957 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501974 4957 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501980 4957 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501985 4957 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501991 4957 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.501997 4957 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.502006 4957 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.502440 4957 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.505500 4957 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.505593 4957 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.506237 4957 server.go:997] "Starting client certificate rotation" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.506264 4957 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.506496 4957 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-24 21:14:16.821849311 +0000 UTC Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.506637 4957 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 447h34m56.31521809s for next certificate rotation Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.512668 4957 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.514652 4957 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.526355 4957 log.go:25] "Validated CRI v1 runtime API" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.542319 4957 log.go:25] "Validated CRI v1 image API" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.543775 4957 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.546337 4957 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-05-34-15-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.546384 4957 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.573152 4957 manager.go:217] Machine: {Timestamp:2025-12-06 05:39:20.570896803 +0000 UTC m=+0.221164505 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:15ddb378-9b6e-4ca1-8e52-89f797d2e2f0 BootID:b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:31:04:a7 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:31:04:a7 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:14:da:3a Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c7:b0:46 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:dc:ad:83 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:da:aa:41 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:e5:d8:90:7e:3b Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8a:f5:da:70:da:4b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.573617 4957 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.573956 4957 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.574452 4957 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.574736 4957 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.574784 4957 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.575172 4957 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.575191 4957 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.575507 4957 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.575559 4957 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.575934 4957 state_mem.go:36] "Initialized new in-memory state store" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.576096 4957 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.576962 4957 kubelet.go:418] "Attempting to sync node with API server" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.577005 4957 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.577057 4957 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.577088 4957 kubelet.go:324] "Adding apiserver pod source" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.577112 4957 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.579823 4957 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.580118 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.580304 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.580162 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.580580 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.580489 4957 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.582300 4957 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583032 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583134 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583216 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583281 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583357 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583422 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583484 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583553 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583624 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583696 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583768 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.583850 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.584172 4957 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.584738 4957 server.go:1280] "Started kubelet" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.590806 4957 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.585373 4957 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.585212 4957 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.591771 4957 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 05:39:20 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.595224 4957 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.595694 4957 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.595901 4957 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.598984 4957 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 05:21:51.600075167 +0000 UTC Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.599223 4957 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 191h42m31.000862557s for next certificate rotation Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.600213 4957 server.go:460] "Adding debug handlers to kubelet server" Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.600324 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.600387 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.601704 4957 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.601724 4957 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.601863 4957 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.601987 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="200ms" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.601775 4957 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.151:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e89bfbd5f3e59 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:39:20.584707673 +0000 UTC m=+0.234975315,LastTimestamp:2025-12-06 05:39:20.584707673 +0000 UTC m=+0.234975315,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607497 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607537 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607549 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607560 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607570 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607581 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607592 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607603 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607619 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607629 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607638 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607647 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607658 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607669 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607679 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607723 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607735 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607746 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607756 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607766 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607777 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607789 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607800 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607810 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607821 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607849 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607867 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607878 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607890 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607900 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607910 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607919 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607931 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607940 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607950 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607960 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607970 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607982 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.607992 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608003 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608014 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608024 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608034 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608047 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608057 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608066 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608076 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608086 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608095 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608105 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608118 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608129 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608142 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608156 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608156 4957 factory.go:55] Registering systemd factory Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608199 4957 factory.go:221] Registration of the systemd container factory successfully Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608165 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608262 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608273 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608283 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608292 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608304 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608314 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608324 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608334 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608343 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608353 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608362 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608372 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608381 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608390 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608400 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608408 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608418 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608427 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608436 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608445 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608455 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608466 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608476 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608487 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608496 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608505 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608515 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608524 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608533 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608543 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608552 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608562 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608572 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608582 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608593 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608605 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608617 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608628 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608637 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608647 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608657 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608668 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608679 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608688 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608697 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608704 4957 factory.go:153] Registering CRI-O factory Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608742 4957 factory.go:221] Registration of the crio container factory successfully Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609108 4957 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609195 4957 factory.go:103] Registering Raw factory Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609230 4957 manager.go:1196] Started watching for new ooms in manager Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.608742 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609344 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609426 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609461 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609513 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609551 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609588 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609624 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609656 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609689 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609719 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609752 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609781 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609819 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609887 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.609919 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610144 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610172 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610200 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610229 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610255 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610287 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610318 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610348 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610375 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610405 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610434 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610463 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610490 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610525 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610551 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610581 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610608 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610638 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610665 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610691 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610721 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610747 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610776 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610802 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.610861 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611084 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611113 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611142 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611168 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611197 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611297 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611869 4957 manager.go:319] Starting recovery of all containers Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.611950 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612008 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612030 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612049 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612064 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612076 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612088 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612103 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612115 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612129 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612141 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612155 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612167 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612181 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612200 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612219 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612233 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612248 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612261 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612274 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612287 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612300 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612312 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612323 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612337 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612353 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612365 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612377 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612388 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612399 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612412 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612425 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612436 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612449 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612462 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612479 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612490 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612502 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612514 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612527 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612539 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612550 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.612563 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613277 4957 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613306 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613321 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613338 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613352 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613365 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613378 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613391 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613406 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613422 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613438 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613453 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613468 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613481 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613495 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613508 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613529 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613543 4957 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613554 4957 reconstruct.go:97] "Volume reconstruction finished" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.613562 4957 reconciler.go:26] "Reconciler: start to sync state" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.638898 4957 manager.go:324] Recovery completed Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.649698 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.651662 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.651712 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.651726 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.652951 4957 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.652970 4957 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.652999 4957 state_mem.go:36] "Initialized new in-memory state store" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.658070 4957 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.660621 4957 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.660691 4957 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.661006 4957 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.661102 4957 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 05:39:20 crc kubenswrapper[4957]: W1206 05:39:20.662928 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.663030 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.663138 4957 policy_none.go:49] "None policy: Start" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.664467 4957 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.664502 4957 state_mem.go:35] "Initializing new in-memory state store" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.696748 4957 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.722976 4957 manager.go:334] "Starting Device Plugin manager" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.723055 4957 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.723074 4957 server.go:79] "Starting device plugin registration server" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.723590 4957 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.723610 4957 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.723885 4957 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.723981 4957 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.723994 4957 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.733722 4957 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.761764 4957 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.761885 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.762788 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.762823 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.762852 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.763035 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.763440 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.763509 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.763948 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.763983 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.763995 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764114 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764217 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764251 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764526 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764543 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764553 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764885 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.764992 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.765010 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.765020 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.765024 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.765097 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.765188 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.765320 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.765372 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766368 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766389 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766397 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766410 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766436 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766449 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766552 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766660 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.766687 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767446 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767493 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767511 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767637 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767674 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767687 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767810 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.767863 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.768686 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.768707 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.768716 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.802862 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="400ms" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816519 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816570 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816603 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816628 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816674 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816717 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816763 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816792 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816815 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816859 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816894 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816910 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816925 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816954 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.816971 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.824513 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.827658 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.827799 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.827887 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.828600 4957 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:39:20 crc kubenswrapper[4957]: E1206 05:39:20.828992 4957 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.151:6443: connect: connection refused" node="crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918508 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918589 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918622 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918657 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918689 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918718 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918748 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918781 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918813 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918877 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918909 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918939 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918968 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.918998 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.919028 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.919744 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.919903 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920005 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920070 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920121 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920171 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920219 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920269 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920321 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920368 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920416 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920468 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920515 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920568 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:20 crc kubenswrapper[4957]: I1206 05:39:20.920686 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.029127 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.030933 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.030998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.031021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.031066 4957 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:39:21 crc kubenswrapper[4957]: E1206 05:39:21.031750 4957 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.151:6443: connect: connection refused" node="crc" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.113784 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.137748 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:21 crc kubenswrapper[4957]: W1206 05:39:21.164993 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9a4f800228b7a75821e1dffba52e27119bf43b095558ece4548f9bffd3862342 WatchSource:0}: Error finding container 9a4f800228b7a75821e1dffba52e27119bf43b095558ece4548f9bffd3862342: Status 404 returned error can't find the container with id 9a4f800228b7a75821e1dffba52e27119bf43b095558ece4548f9bffd3862342 Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.175076 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.199297 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:21 crc kubenswrapper[4957]: W1206 05:39:21.202518 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-8df4309463920d1d7d69fcba90e213c132799b2b748de510d566b0efb67e3755 WatchSource:0}: Error finding container 8df4309463920d1d7d69fcba90e213c132799b2b748de510d566b0efb67e3755: Status 404 returned error can't find the container with id 8df4309463920d1d7d69fcba90e213c132799b2b748de510d566b0efb67e3755 Dec 06 05:39:21 crc kubenswrapper[4957]: E1206 05:39:21.203731 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="800ms" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.210777 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 05:39:21 crc kubenswrapper[4957]: W1206 05:39:21.215535 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-dc7c037c246ff82e7280d7d96ab79f2b7560229f6f94d9d2eceeab4908832f84 WatchSource:0}: Error finding container dc7c037c246ff82e7280d7d96ab79f2b7560229f6f94d9d2eceeab4908832f84: Status 404 returned error can't find the container with id dc7c037c246ff82e7280d7d96ab79f2b7560229f6f94d9d2eceeab4908832f84 Dec 06 05:39:21 crc kubenswrapper[4957]: W1206 05:39:21.230533 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-602869ad982ed53a112b858945b3eafc0cb755783ac1dc0cf746b4b15f584a07 WatchSource:0}: Error finding container 602869ad982ed53a112b858945b3eafc0cb755783ac1dc0cf746b4b15f584a07: Status 404 returned error can't find the container with id 602869ad982ed53a112b858945b3eafc0cb755783ac1dc0cf746b4b15f584a07 Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.432457 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.436577 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.436625 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.436636 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.436665 4957 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:39:21 crc kubenswrapper[4957]: E1206 05:39:21.437132 4957 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.151:6443: connect: connection refused" node="crc" Dec 06 05:39:21 crc kubenswrapper[4957]: W1206 05:39:21.534443 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:21 crc kubenswrapper[4957]: E1206 05:39:21.534556 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.592547 4957 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.672026 4957 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae" exitCode=0 Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.672124 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.672214 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"602869ad982ed53a112b858945b3eafc0cb755783ac1dc0cf746b4b15f584a07"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.672378 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.674633 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.674698 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.674720 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.676236 4957 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c" exitCode=0 Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.676294 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.676348 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dc7c037c246ff82e7280d7d96ab79f2b7560229f6f94d9d2eceeab4908832f84"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.676479 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.678168 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.678281 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.678339 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.678756 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.678788 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8df4309463920d1d7d69fcba90e213c132799b2b748de510d566b0efb67e3755"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.680493 4957 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f" exitCode=0 Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.680561 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.680593 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a4f800228b7a75821e1dffba52e27119bf43b095558ece4548f9bffd3862342"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.680689 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.681695 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.681728 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.681740 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.682652 4957 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b516033a1602d6c7ee6846ba9f60052d8b6c1c0301b75ebbe81a7b3be27455f9" exitCode=0 Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.682704 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b516033a1602d6c7ee6846ba9f60052d8b6c1c0301b75ebbe81a7b3be27455f9"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.682756 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"275b8fccb5f373e3d7e503265c45d0202e49b92d095d001132bf0dedc3df06e7"} Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.682969 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.684323 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.684365 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.684389 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.686223 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.687137 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.687182 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:21 crc kubenswrapper[4957]: I1206 05:39:21.687200 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:21 crc kubenswrapper[4957]: W1206 05:39:21.905146 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:21 crc kubenswrapper[4957]: E1206 05:39:21.905237 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:22 crc kubenswrapper[4957]: E1206 05:39:22.004959 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="1.6s" Dec 06 05:39:22 crc kubenswrapper[4957]: W1206 05:39:22.171997 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:22 crc kubenswrapper[4957]: E1206 05:39:22.172114 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:22 crc kubenswrapper[4957]: W1206 05:39:22.181039 4957 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.151:6443: connect: connection refused Dec 06 05:39:22 crc kubenswrapper[4957]: E1206 05:39:22.181169 4957 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.151:6443: connect: connection refused" logger="UnhandledError" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.237725 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.239957 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.240003 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.240811 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.240864 4957 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:39:22 crc kubenswrapper[4957]: E1206 05:39:22.242320 4957 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.151:6443: connect: connection refused" node="crc" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.688137 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.688212 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.688231 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.688247 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.691044 4957 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d7cd4d9873f19a4bf74bf4423c6899a286bf139dd69d60b9ee469b1557fdc428" exitCode=0 Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.691105 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d7cd4d9873f19a4bf74bf4423c6899a286bf139dd69d60b9ee469b1557fdc428"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.691225 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.694010 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.694036 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.694045 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.695825 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e9af17c49e0732a21c5ccc770ca8367336ae1b9c225bafebb180a403a5d002d5"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.695907 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.696737 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.696780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.696796 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.704241 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.704266 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.704277 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.704343 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.704897 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.704925 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.704937 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.707715 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.707743 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.707757 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b"} Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.707824 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.709299 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.709323 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:22 crc kubenswrapper[4957]: I1206 05:39:22.709334 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.699652 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.715307 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9"} Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.715384 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.721918 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.722059 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.722145 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.726270 4957 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c0438adc14a024c7ee6ec65013516ce96a1552819370fc9fb0e830f3e61019e8" exitCode=0 Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.726377 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c0438adc14a024c7ee6ec65013516ce96a1552819370fc9fb0e830f3e61019e8"} Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.726446 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.726662 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.727660 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.727710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.727733 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.729178 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.729222 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.729242 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.842676 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.844047 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.844095 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.844108 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:23 crc kubenswrapper[4957]: I1206 05:39:23.844140 4957 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.737194 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9434df9fc2bffa7a660a7703be5a27dbd7c4190d4b4eba0b989601ba11dd8c28"} Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.737262 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c3d835f8ae3403436d54eee9a7b4fbb4e7f080c3601b8f42bbc9dd74b972015e"} Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.737287 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f9a274615fdfcff2b2ca7206129ee58d99c4e1ca8009ff19fd2bf1e8fb37129"} Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.737306 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b35f28af4d11fd50c5c287d92688bcdd58b4183fab47cafcf4c2ed8f746d19ea"} Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.737269 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.737381 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.737393 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.738699 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.738761 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.738789 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.738889 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.738929 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:24 crc kubenswrapper[4957]: I1206 05:39:24.738954 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:25 crc kubenswrapper[4957]: I1206 05:39:25.745628 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9c496a594758f3caf99a222f181c66da51ac1f77bac9c533ade93a251f0d24bb"} Dec 06 05:39:25 crc kubenswrapper[4957]: I1206 05:39:25.745760 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:25 crc kubenswrapper[4957]: I1206 05:39:25.746487 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:25 crc kubenswrapper[4957]: I1206 05:39:25.746512 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:25 crc kubenswrapper[4957]: I1206 05:39:25.746521 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.400054 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.400295 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.402040 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.402074 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.402123 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.699943 4957 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.700026 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.748133 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.749432 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.749484 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:26 crc kubenswrapper[4957]: I1206 05:39:26.749501 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:27 crc kubenswrapper[4957]: I1206 05:39:27.886953 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:27 crc kubenswrapper[4957]: I1206 05:39:27.887273 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:39:27 crc kubenswrapper[4957]: I1206 05:39:27.887346 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:27 crc kubenswrapper[4957]: I1206 05:39:27.889050 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:27 crc kubenswrapper[4957]: I1206 05:39:27.889112 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:27 crc kubenswrapper[4957]: I1206 05:39:27.889131 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:28 crc kubenswrapper[4957]: I1206 05:39:28.469434 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:28 crc kubenswrapper[4957]: I1206 05:39:28.753790 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:39:28 crc kubenswrapper[4957]: I1206 05:39:28.753928 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:28 crc kubenswrapper[4957]: I1206 05:39:28.755487 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:28 crc kubenswrapper[4957]: I1206 05:39:28.755561 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:28 crc kubenswrapper[4957]: I1206 05:39:28.755583 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.002043 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.002462 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.004634 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.004707 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.004726 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.762214 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.762455 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.764585 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.764642 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.764661 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:29 crc kubenswrapper[4957]: I1206 05:39:29.769870 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.490640 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.491125 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.495319 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.495384 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.495434 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:30 crc kubenswrapper[4957]: E1206 05:39:30.734348 4957 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.760475 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.760602 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.761928 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.761994 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:30 crc kubenswrapper[4957]: I1206 05:39:30.762014 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.382727 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.763975 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.765325 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.765441 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.765468 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.899692 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.899985 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.901434 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.901478 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:31 crc kubenswrapper[4957]: I1206 05:39:31.901487 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:32 crc kubenswrapper[4957]: I1206 05:39:32.592032 4957 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 06 05:39:32 crc kubenswrapper[4957]: I1206 05:39:32.767084 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:32 crc kubenswrapper[4957]: I1206 05:39:32.768526 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:32 crc kubenswrapper[4957]: I1206 05:39:32.768579 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:32 crc kubenswrapper[4957]: I1206 05:39:32.768597 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:33 crc kubenswrapper[4957]: I1206 05:39:33.215107 4957 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 05:39:33 crc kubenswrapper[4957]: I1206 05:39:33.215196 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 05:39:33 crc kubenswrapper[4957]: I1206 05:39:33.224314 4957 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 05:39:33 crc kubenswrapper[4957]: I1206 05:39:33.224386 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.738793 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.739062 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.740319 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.740377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.740396 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.760801 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.777074 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.778277 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.778308 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.778317 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:35 crc kubenswrapper[4957]: I1206 05:39:35.792715 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 05:39:36 crc kubenswrapper[4957]: I1206 05:39:36.700582 4957 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 05:39:36 crc kubenswrapper[4957]: I1206 05:39:36.701006 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 05:39:36 crc kubenswrapper[4957]: I1206 05:39:36.779645 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:36 crc kubenswrapper[4957]: I1206 05:39:36.780823 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:36 crc kubenswrapper[4957]: I1206 05:39:36.780883 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:36 crc kubenswrapper[4957]: I1206 05:39:36.780901 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:37 crc kubenswrapper[4957]: I1206 05:39:37.892033 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:37 crc kubenswrapper[4957]: I1206 05:39:37.892350 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:37 crc kubenswrapper[4957]: I1206 05:39:37.894558 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:37 crc kubenswrapper[4957]: I1206 05:39:37.894685 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:37 crc kubenswrapper[4957]: I1206 05:39:37.894759 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:37 crc kubenswrapper[4957]: I1206 05:39:37.900505 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.185788 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.188129 4957 trace.go:236] Trace[906714804]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:39:23.943) (total time: 14244ms): Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[906714804]: ---"Objects listed" error: 14244ms (05:39:38.188) Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[906714804]: [14.244187916s] [14.244187916s] END Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.188481 4957 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.189590 4957 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.609158 4957 trace.go:236] Trace[1755182393]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:39:25.282) (total time: 13326ms): Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[1755182393]: ---"Objects listed" error: 13326ms (05:39:38.608) Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[1755182393]: [13.326226642s] [13.326226642s] END Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.609232 4957 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.612327 4957 apiserver.go:52] "Watching apiserver" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.614294 4957 trace.go:236] Trace[2139641348]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:39:25.071) (total time: 13542ms): Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[2139641348]: ---"Objects listed" error: 13540ms (05:39:38.612) Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[2139641348]: [13.54240623s] [13.54240623s] END Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.614388 4957 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.616480 4957 trace.go:236] Trace[1355955158]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 05:39:23.266) (total time: 15349ms): Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[1355955158]: ---"Objects listed" error: 15349ms (05:39:38.615) Dec 06 05:39:38 crc kubenswrapper[4957]: Trace[1355955158]: [15.349484036s] [15.349484036s] END Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.616521 4957 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.623045 4957 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.744175 4957 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.744570 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.745079 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.745195 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.745256 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.745379 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.745484 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.745590 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.746074 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.746482 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.746540 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.748025 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.748280 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.748336 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.749600 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.749884 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.750538 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.750606 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.751725 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.752039 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.789539 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.800310 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.802956 4957 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.814858 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825356 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825410 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825433 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825455 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825475 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825494 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825545 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825564 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825581 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825621 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825664 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825686 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825734 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825762 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825785 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825802 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825824 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825868 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825888 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825908 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825927 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825911 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825948 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.825914 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826001 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826023 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826049 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826069 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826088 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826105 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826124 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826142 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826143 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826162 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826240 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826270 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826298 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826324 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826347 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826374 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826402 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826423 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826445 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826473 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826497 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826519 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826542 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826569 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826576 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826592 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826567 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826622 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826662 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826688 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826705 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826723 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826740 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826757 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826773 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826789 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826806 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.826823 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827061 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827084 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827099 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827158 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827296 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827343 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827361 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827429 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827466 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827536 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827571 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827624 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827677 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827732 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827835 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827898 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827913 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827940 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828058 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828300 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828436 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.827706 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828498 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828522 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828541 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828558 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828579 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828602 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828621 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828639 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828659 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828714 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828734 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828755 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828777 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828798 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828818 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828840 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828875 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828894 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828913 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828934 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828952 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828970 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.828988 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829007 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829027 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829050 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829071 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829090 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829131 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829149 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829167 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829184 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829210 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829234 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829293 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829318 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829343 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829362 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829390 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829409 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829431 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829453 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829472 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829490 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829509 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829526 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829546 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829567 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829587 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829603 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829619 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829636 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829653 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829671 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829688 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829704 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829720 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829738 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829757 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829774 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829790 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829810 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.829827 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830691 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830724 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830744 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830762 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830780 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830798 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830815 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830837 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830938 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830958 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.830980 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831000 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831018 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831036 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831053 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831070 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831086 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831103 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831121 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831141 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831161 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831179 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831197 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831217 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831234 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831256 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831274 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831292 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831310 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831329 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831345 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831361 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831378 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831396 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831413 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831430 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831481 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831498 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831517 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831533 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831556 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831583 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831604 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831628 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831650 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831669 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831687 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831704 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831721 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831738 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831754 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831773 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831792 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831810 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831827 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831868 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831886 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831902 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831924 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831941 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831957 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831975 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.831990 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832006 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832025 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832042 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832061 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832077 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832095 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832116 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832157 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832185 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832202 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832219 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832266 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832293 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832311 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832331 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832350 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832375 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832397 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832420 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832450 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832467 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832491 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832512 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832531 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832550 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832810 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832827 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832857 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832876 4957 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832894 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832908 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832921 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832933 4957 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832948 4957 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832962 4957 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832973 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832985 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.832998 4957 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833013 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833028 4957 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833042 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833055 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833065 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833074 4957 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833030 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833098 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833184 4957 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833208 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833228 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833245 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833266 4957 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.833280 4957 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.834028 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.836283 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.836412 4957 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52212->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.836451 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52212->192.168.126.11:17697: read: connection reset by peer" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.836718 4957 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.836998 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.841972 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.842537 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.842879 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.843042 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.843135 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.843599 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.843924 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.844215 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.845206 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.845570 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.845984 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.846249 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.846264 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.846355 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.846658 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:39:39.346628482 +0000 UTC m=+18.996896114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.847014 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.847045 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.847369 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.848031 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.848472 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.848653 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.848794 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.848999 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.849278 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.849632 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.848249 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.851161 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.851236 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.851540 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.851758 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.852101 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.852314 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.852453 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.853098 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.854520 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.855491 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.857574 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.860477 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.860889 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.866307 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.866443 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.867040 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.867178 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.872411 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.872516 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:39.372425 +0000 UTC m=+19.022692632 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.867307 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.872696 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:39.372687557 +0000 UTC m=+19.022955189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.867598 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.867363 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.868460 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.868805 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.869401 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.869428 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.869702 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.869718 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.870281 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.870610 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.870940 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.871438 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.871859 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.872231 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.872914 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.873458 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.873905 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.874432 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.874674 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.873979 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.875011 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.875354 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.878241 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.879291 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.879686 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.879977 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.880307 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.880865 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.881470 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.881749 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.881906 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.881991 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.882142 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.882620 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.882697 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.882752 4957 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.883205 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.883339 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.883579 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.884330 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.884585 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.884685 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.884808 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.885314 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.887332 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.888035 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.888522 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.888624 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.888808 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.888868 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.889110 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.889386 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.889527 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.889893 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.890039 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.891071 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.891203 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.891223 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.891447 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.891519 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.891557 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.891745 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.892044 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.892809 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.893219 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.893339 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.893786 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.895600 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.896503 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.896543 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.889578 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.896926 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.897050 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.897110 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.897328 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.897402 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.897474 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.898608 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.900387 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.900739 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.903008 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.903083 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.904148 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.904157 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.907920 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.908989 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.909028 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.909045 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.909136 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:39.409110294 +0000 UTC m=+19.059377926 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.910111 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.910150 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.910173 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:38 crc kubenswrapper[4957]: E1206 05:39:38.910249 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:39.410216424 +0000 UTC m=+19.060484046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.917870 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.919342 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.919656 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.919660 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.919708 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.919909 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.920305 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.920563 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.920728 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.920876 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.922043 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.923521 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.924078 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.924170 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.924576 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.924736 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.924796 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.925268 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.925895 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.926301 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.926463 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.926478 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.926671 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.931027 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.931602 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.931739 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.932206 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.934727 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935388 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935728 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935775 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935897 4957 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935919 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935931 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935945 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935955 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935964 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935974 4957 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935985 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.935994 4957 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936004 4957 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936016 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936018 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936028 4957 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936063 4957 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936067 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936078 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936089 4957 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936098 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936108 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936117 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936127 4957 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936137 4957 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936146 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936157 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936167 4957 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936177 4957 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936218 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936228 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936237 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936243 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936247 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936265 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936275 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936286 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936299 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936310 4957 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936319 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936331 4957 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936344 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936358 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936370 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936382 4957 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936392 4957 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936401 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936411 4957 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936421 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936433 4957 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936445 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936459 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936472 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936482 4957 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936491 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936501 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936509 4957 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936505 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936522 4957 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936612 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936626 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936640 4957 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936652 4957 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936666 4957 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936677 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936688 4957 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936698 4957 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936708 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936719 4957 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936729 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936739 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936748 4957 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936759 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936769 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936782 4957 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936792 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936801 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936811 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936823 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936857 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936868 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936878 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936888 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936899 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936909 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936919 4957 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936931 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936940 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936952 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936963 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936974 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936985 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.936997 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937007 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937018 4957 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937069 4957 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937081 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937091 4957 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937101 4957 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937112 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937122 4957 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937131 4957 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937140 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937149 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937159 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937169 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937178 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937190 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937201 4957 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937210 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937220 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937229 4957 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937238 4957 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937248 4957 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937257 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937267 4957 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937279 4957 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937291 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937300 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937309 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937317 4957 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937327 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937337 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937348 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937360 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937369 4957 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937378 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937388 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937398 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937409 4957 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937419 4957 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937427 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937437 4957 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937447 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937456 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937467 4957 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937476 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937487 4957 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937496 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937506 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937515 4957 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937525 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937534 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937543 4957 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937553 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937562 4957 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937571 4957 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937581 4957 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937590 4957 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937601 4957 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937609 4957 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937621 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937629 4957 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937638 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937884 4957 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.937895 4957 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.938928 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.939119 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.939354 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.939664 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.941119 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.941807 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.941982 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.942165 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.942754 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.942764 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.942910 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.943382 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.943419 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.943629 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.944219 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.944223 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.945289 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.956958 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.960156 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.966736 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:38 crc kubenswrapper[4957]: I1206 05:39:38.970140 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.020323 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-69r8l"] Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.020638 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.023961 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.024068 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-g9hn4"] Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.024650 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.024736 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.024799 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.024995 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.026583 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hlm8s"] Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.027719 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.032233 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.032233 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.032406 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-f2z46"] Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.032366 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.032960 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.033105 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-78cpc"] Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.033623 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.037434 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.037870 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.037971 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038068 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038158 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038297 4957 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038324 4957 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038337 4957 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038349 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038360 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038383 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038394 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038407 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038418 4957 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038429 4957 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038440 4957 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038451 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038462 4957 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038481 4957 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038494 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038505 4957 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038520 4957 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038534 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038554 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.038576 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.040401 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.040429 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.040577 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.040704 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.040767 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.040967 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.042094 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.042573 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.044687 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.046481 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.059103 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.063818 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.068887 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 05:39:39 crc kubenswrapper[4957]: W1206 05:39:39.072331 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-668e2c95d52a5af20a4a903aa222c923f9a9546084e8ad9ea9f04d398e8506be WatchSource:0}: Error finding container 668e2c95d52a5af20a4a903aa222c923f9a9546084e8ad9ea9f04d398e8506be: Status 404 returned error can't find the container with id 668e2c95d52a5af20a4a903aa222c923f9a9546084e8ad9ea9f04d398e8506be Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.074797 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.078743 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: W1206 05:39:39.083579 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-a052d37c8a430b6fbd9eb42c01bc8b992ab2e4cf5c8798fae14d9addec1d6775 WatchSource:0}: Error finding container a052d37c8a430b6fbd9eb42c01bc8b992ab2e4cf5c8798fae14d9addec1d6775: Status 404 returned error can't find the container with id a052d37c8a430b6fbd9eb42c01bc8b992ab2e4cf5c8798fae14d9addec1d6775 Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.095589 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.106915 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.122441 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.134548 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141504 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-daemon-config\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141565 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-ovn\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141592 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-k8s-cni-cncf-io\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141622 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-var-lib-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141648 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-env-overrides\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141673 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-conf-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141695 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-multus-certs\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141719 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-slash\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141741 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-netns\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141765 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-node-log\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141789 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-log-socket\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141818 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnxw9\" (UniqueName: \"kubernetes.io/projected/446a0886-35d8-4574-9ccd-6b38f6df37bc-kube-api-access-nnxw9\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.141870 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-ovn-kubernetes\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.142023 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c3d0bce5-6056-4247-8eef-23ed810df082-hosts-file\") pod \"node-resolver-78cpc\" (UID: \"c3d0bce5-6056-4247-8eef-23ed810df082\") " pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145202 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1384ce42-5e9a-4d6a-a299-f564d1494b0a-cni-binary-copy\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145280 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cnibin\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145309 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5km2\" (UniqueName: \"kubernetes.io/projected/a8f0915e-a437-4104-8c8c-c7c10fbf8255-kube-api-access-f5km2\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145338 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-etc-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145363 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-config\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145398 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-systemd\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145476 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-cni-bin\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145504 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-etc-kubernetes\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145533 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-system-cni-dir\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145559 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-os-release\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145584 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-netd\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145614 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-system-cni-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145638 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-kubelet\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145669 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145798 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlzxt\" (UniqueName: \"kubernetes.io/projected/1384ce42-5e9a-4d6a-a299-f564d1494b0a-kube-api-access-dlzxt\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.145890 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-systemd-units\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146001 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-os-release\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146064 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-kubelet\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146099 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-tuning-conf-dir\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146140 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-cni-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146178 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-netns\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146204 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-cni-multus\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146360 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-hostroot\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146409 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/446a0886-35d8-4574-9ccd-6b38f6df37bc-rootfs\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146460 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146491 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bpjx\" (UniqueName: \"kubernetes.io/projected/fa635ecb-5324-449e-a8f3-7a9bfdca7064-kube-api-access-6bpjx\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146542 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-cnibin\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146579 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cni-binary-copy\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146610 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146642 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-bin\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146671 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-script-lib\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146700 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/446a0886-35d8-4574-9ccd-6b38f6df37bc-mcd-auth-proxy-config\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146752 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovn-node-metrics-cert\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146788 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4skp\" (UniqueName: \"kubernetes.io/projected/c3d0bce5-6056-4247-8eef-23ed810df082-kube-api-access-q4skp\") pod \"node-resolver-78cpc\" (UID: \"c3d0bce5-6056-4247-8eef-23ed810df082\") " pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146871 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-socket-dir-parent\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.146921 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/446a0886-35d8-4574-9ccd-6b38f6df37bc-proxy-tls\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.147172 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.160564 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.174544 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.185539 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.195497 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.204990 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.213896 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.224616 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.238044 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.248239 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-socket-dir-parent\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.248959 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/446a0886-35d8-4574-9ccd-6b38f6df37bc-proxy-tls\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249029 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4skp\" (UniqueName: \"kubernetes.io/projected/c3d0bce5-6056-4247-8eef-23ed810df082-kube-api-access-q4skp\") pod \"node-resolver-78cpc\" (UID: \"c3d0bce5-6056-4247-8eef-23ed810df082\") " pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249058 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-ovn\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249136 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-daemon-config\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249172 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-k8s-cni-cncf-io\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249201 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-var-lib-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249231 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-env-overrides\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249257 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-multus-certs\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249286 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-slash\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249310 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-netns\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249335 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-node-log\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249361 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-log-socket\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249391 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-conf-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249416 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-ovn-kubernetes\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249447 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnxw9\" (UniqueName: \"kubernetes.io/projected/446a0886-35d8-4574-9ccd-6b38f6df37bc-kube-api-access-nnxw9\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249474 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1384ce42-5e9a-4d6a-a299-f564d1494b0a-cni-binary-copy\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249516 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cnibin\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249546 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5km2\" (UniqueName: \"kubernetes.io/projected/a8f0915e-a437-4104-8c8c-c7c10fbf8255-kube-api-access-f5km2\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249571 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-etc-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249597 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-config\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249627 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c3d0bce5-6056-4247-8eef-23ed810df082-hosts-file\") pod \"node-resolver-78cpc\" (UID: \"c3d0bce5-6056-4247-8eef-23ed810df082\") " pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249653 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-systemd\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249683 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-cni-bin\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249708 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-etc-kubernetes\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249734 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-system-cni-dir\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249757 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-os-release\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249780 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-netd\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249837 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249875 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-system-cni-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249892 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-kubelet\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249910 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlzxt\" (UniqueName: \"kubernetes.io/projected/1384ce42-5e9a-4d6a-a299-f564d1494b0a-kube-api-access-dlzxt\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249939 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-systemd-units\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249963 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-kubelet\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.249994 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-tuning-conf-dir\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250015 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-os-release\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250035 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-cni-multus\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250052 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-hostroot\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250067 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/446a0886-35d8-4574-9ccd-6b38f6df37bc-rootfs\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250088 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-cni-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250105 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-netns\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250122 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250139 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bpjx\" (UniqueName: \"kubernetes.io/projected/fa635ecb-5324-449e-a8f3-7a9bfdca7064-kube-api-access-6bpjx\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250161 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-cnibin\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250224 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cni-binary-copy\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250242 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250259 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-bin\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250279 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-script-lib\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250294 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovn-node-metrics-cert\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.250309 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/446a0886-35d8-4574-9ccd-6b38f6df37bc-mcd-auth-proxy-config\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.253420 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/446a0886-35d8-4574-9ccd-6b38f6df37bc-mcd-auth-proxy-config\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.253766 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-socket-dir-parent\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.253875 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-log-socket\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.253896 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-etc-kubernetes\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.253976 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-conf-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.254014 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-ovn-kubernetes\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.254451 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-system-cni-dir\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.254967 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-os-release\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255014 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-netd\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255049 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255096 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-system-cni-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255123 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-kubelet\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255301 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1384ce42-5e9a-4d6a-a299-f564d1494b0a-cni-binary-copy\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255360 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cnibin\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255452 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-systemd-units\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255498 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-kubelet\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.255541 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-etc-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.256088 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-config\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.256273 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c3d0bce5-6056-4247-8eef-23ed810df082-hosts-file\") pod \"node-resolver-78cpc\" (UID: \"c3d0bce5-6056-4247-8eef-23ed810df082\") " pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.256311 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-systemd\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.256344 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-cni-bin\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.256794 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-ovn\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257245 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-daemon-config\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257290 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-k8s-cni-cncf-io\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257317 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-var-lib-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257634 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-env-overrides\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257680 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-multus-certs\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257714 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-slash\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257741 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-netns\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257768 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-node-log\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257787 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-os-release\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257915 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-cnibin\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257947 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-var-lib-cni-multus\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.257982 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-hostroot\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.258012 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/446a0886-35d8-4574-9ccd-6b38f6df37bc-rootfs\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.258196 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-multus-cni-dir\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.258239 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-bin\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.258268 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-openvswitch\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.258388 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1384ce42-5e9a-4d6a-a299-f564d1494b0a-host-run-netns\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.258815 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-script-lib\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.258900 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cni-binary-copy\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.259269 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a8f0915e-a437-4104-8c8c-c7c10fbf8255-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.259712 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/446a0886-35d8-4574-9ccd-6b38f6df37bc-proxy-tls\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.261381 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.263586 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovn-node-metrics-cert\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.270111 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.270589 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnxw9\" (UniqueName: \"kubernetes.io/projected/446a0886-35d8-4574-9ccd-6b38f6df37bc-kube-api-access-nnxw9\") pod \"machine-config-daemon-f2z46\" (UID: \"446a0886-35d8-4574-9ccd-6b38f6df37bc\") " pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.273459 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5km2\" (UniqueName: \"kubernetes.io/projected/a8f0915e-a437-4104-8c8c-c7c10fbf8255-kube-api-access-f5km2\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.273762 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlzxt\" (UniqueName: \"kubernetes.io/projected/1384ce42-5e9a-4d6a-a299-f564d1494b0a-kube-api-access-dlzxt\") pod \"multus-69r8l\" (UID: \"1384ce42-5e9a-4d6a-a299-f564d1494b0a\") " pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.274457 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4skp\" (UniqueName: \"kubernetes.io/projected/c3d0bce5-6056-4247-8eef-23ed810df082-kube-api-access-q4skp\") pod \"node-resolver-78cpc\" (UID: \"c3d0bce5-6056-4247-8eef-23ed810df082\") " pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.279726 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bpjx\" (UniqueName: \"kubernetes.io/projected/fa635ecb-5324-449e-a8f3-7a9bfdca7064-kube-api-access-6bpjx\") pod \"ovnkube-node-hlm8s\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.281661 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.292914 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.307234 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a8f0915e-a437-4104-8c8c-c7c10fbf8255-tuning-conf-dir\") pod \"multus-additional-cni-plugins-g9hn4\" (UID: \"a8f0915e-a437-4104-8c8c-c7c10fbf8255\") " pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.341823 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-69r8l" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.350324 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.351368 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.351588 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:39:40.351551433 +0000 UTC m=+20.001819085 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.357365 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:39 crc kubenswrapper[4957]: W1206 05:39:39.358770 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1384ce42_5e9a_4d6a_a299_f564d1494b0a.slice/crio-d54f075f3a86e7d5dca58779494a8b6d855d74ab1d345215af956bdc6fa6beb6 WatchSource:0}: Error finding container d54f075f3a86e7d5dca58779494a8b6d855d74ab1d345215af956bdc6fa6beb6: Status 404 returned error can't find the container with id d54f075f3a86e7d5dca58779494a8b6d855d74ab1d345215af956bdc6fa6beb6 Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.365886 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.378163 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-78cpc" Dec 06 05:39:39 crc kubenswrapper[4957]: W1206 05:39:39.396001 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa635ecb_5324_449e_a8f3_7a9bfdca7064.slice/crio-f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 WatchSource:0}: Error finding container f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5: Status 404 returned error can't find the container with id f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 Dec 06 05:39:39 crc kubenswrapper[4957]: W1206 05:39:39.405804 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod446a0886_35d8_4574_9ccd_6b38f6df37bc.slice/crio-f702855a843508ff7fee7c081a6af3de061c76383bd434bc6ad960fb1a1d33f6 WatchSource:0}: Error finding container f702855a843508ff7fee7c081a6af3de061c76383bd434bc6ad960fb1a1d33f6: Status 404 returned error can't find the container with id f702855a843508ff7fee7c081a6af3de061c76383bd434bc6ad960fb1a1d33f6 Dec 06 05:39:39 crc kubenswrapper[4957]: W1206 05:39:39.433000 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3d0bce5_6056_4247_8eef_23ed810df082.slice/crio-50216e55b66cbc34e75dfbdb238cf24d0988bdc4a676fe55a47ac2a21d20afb7 WatchSource:0}: Error finding container 50216e55b66cbc34e75dfbdb238cf24d0988bdc4a676fe55a47ac2a21d20afb7: Status 404 returned error can't find the container with id 50216e55b66cbc34e75dfbdb238cf24d0988bdc4a676fe55a47ac2a21d20afb7 Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.453078 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.453124 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.453162 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.453185 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453340 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453358 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453371 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453425 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:40.453406422 +0000 UTC m=+20.103674044 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453484 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453494 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453502 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453526 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:40.453516765 +0000 UTC m=+20.103784397 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453563 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453585 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:40.453580257 +0000 UTC m=+20.103847889 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453629 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: E1206 05:39:39.453665 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:40.453643808 +0000 UTC m=+20.103911440 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.790413 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-78cpc" event={"ID":"c3d0bce5-6056-4247-8eef-23ed810df082","Type":"ContainerStarted","Data":"4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.790475 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-78cpc" event={"ID":"c3d0bce5-6056-4247-8eef-23ed810df082","Type":"ContainerStarted","Data":"50216e55b66cbc34e75dfbdb238cf24d0988bdc4a676fe55a47ac2a21d20afb7"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.793058 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.793096 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.793106 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"f702855a843508ff7fee7c081a6af3de061c76383bd434bc6ad960fb1a1d33f6"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.794731 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerStarted","Data":"5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.794763 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerStarted","Data":"49d35b2849f14a2e7a3eef37bd631716fd6bf18dcf96ab80dd3b882ae024b751"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.796888 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3bea8da80d9f5e4d7cd2846a1634b19bf29b6ee52881e16c7824d6df5e90274e"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.798382 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.798410 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.798420 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a052d37c8a430b6fbd9eb42c01bc8b992ab2e4cf5c8798fae14d9addec1d6775"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.800493 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.802073 4957 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9" exitCode=255 Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.802126 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.802644 4957 scope.go:117] "RemoveContainer" containerID="1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.804609 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.804633 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"668e2c95d52a5af20a4a903aa222c923f9a9546084e8ad9ea9f04d398e8506be"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.805820 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d" exitCode=0 Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.805886 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.805952 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.807226 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerStarted","Data":"0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.807257 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerStarted","Data":"d54f075f3a86e7d5dca58779494a8b6d855d74ab1d345215af956bdc6fa6beb6"} Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.829620 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.860722 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.882065 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.921167 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.936924 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.958265 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:39 crc kubenswrapper[4957]: I1206 05:39:39.980264 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.000500 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:39Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.014972 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.030548 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.043948 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.058317 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.072963 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.089030 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.112290 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.130154 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.148996 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.166704 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.181381 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.202358 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.219349 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.247990 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.259140 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.268373 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.362301 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.362652 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:39:42.362619068 +0000 UTC m=+22.012886720 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.422162 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.464072 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464223 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.464429 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464525 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:42.464497508 +0000 UTC m=+22.114765140 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.464602 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.464653 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464758 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464859 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464882 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464897 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464784 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464941 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464950 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.464870 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:42.464856888 +0000 UTC m=+22.115124520 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.465137 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:42.465128315 +0000 UTC m=+22.115395947 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.465208 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:42.465200357 +0000 UTC m=+22.115467979 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.662156 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.662753 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.665136 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.667997 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.667206 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.667893 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: E1206 05:39:40.667102 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.669173 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.670058 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.671451 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.672179 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.673422 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.674408 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.675113 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.680166 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.681314 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.684370 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.685599 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.687584 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.688517 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.690706 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.691549 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.692615 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.693183 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.693871 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.694545 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.695162 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.695763 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.697719 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.698712 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.700528 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.701266 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.703301 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.703837 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.704507 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.705142 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.706060 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.706547 4957 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.706655 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.709726 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.711053 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.711615 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.714739 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.715738 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.716457 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.717729 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.719177 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.719831 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.720677 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.721759 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.723835 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.724392 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.725441 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.726074 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.727637 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.728205 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.728525 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.729629 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.730502 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.731126 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.732221 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.732713 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.750519 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.778242 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.797065 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.812478 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.813804 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488"} Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.814156 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.816504 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.818728 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b"} Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.818859 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d"} Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.818925 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4"} Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.818994 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f"} Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.820957 4957 generic.go:334] "Generic (PLEG): container finished" podID="a8f0915e-a437-4104-8c8c-c7c10fbf8255" containerID="5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1" exitCode=0 Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.821006 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerDied","Data":"5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1"} Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.829290 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.842623 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.856920 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.872541 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.891326 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.906644 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.924774 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.946226 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.975087 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:40 crc kubenswrapper[4957]: I1206 05:39:40.988160 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.010127 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.025580 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.045206 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.058253 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.075021 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.088501 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.104082 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.136113 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.391610 4957 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.395937 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.395996 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.396013 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.396221 4957 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.406005 4957 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.406414 4957 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.407873 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.407943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.407963 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.407990 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.408012 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: E1206 05:39:41.431102 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.435906 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.436123 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.436260 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.436402 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.436549 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: E1206 05:39:41.453899 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.458573 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.458609 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.458623 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.458639 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.458651 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: E1206 05:39:41.472476 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.477217 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.477255 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.477267 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.477283 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.477334 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: E1206 05:39:41.491787 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.495806 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.495887 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.495900 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.495923 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.495942 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: E1206 05:39:41.507724 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: E1206 05:39:41.507917 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.510385 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.510430 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.510443 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.510464 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.510477 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.613915 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.613971 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.613991 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.614022 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.614042 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.717964 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.718042 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.718065 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.718096 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.718117 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.821633 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.821690 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.821700 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.821720 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.821731 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.826583 4957 generic.go:334] "Generic (PLEG): container finished" podID="a8f0915e-a437-4104-8c8c-c7c10fbf8255" containerID="3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a" exitCode=0 Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.826720 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerDied","Data":"3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.832929 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.832980 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.851861 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.870961 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.887413 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.904204 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.922895 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.924943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.925025 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.925037 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.925051 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.925184 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:41Z","lastTransitionTime":"2025-12-06T05:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.936965 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.950940 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.968391 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:41 crc kubenswrapper[4957]: I1206 05:39:41.982492 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.006120 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.028308 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.028676 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.028741 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.028806 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.028892 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.039050 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.062392 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.132159 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.132208 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.132224 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.132247 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.132264 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.234391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.234438 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.234451 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.234471 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.234485 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.337634 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.337671 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.337682 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.337695 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.337705 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.384581 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.384765 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:39:46.384732969 +0000 UTC m=+26.035000601 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.441060 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.441102 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.441117 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.441132 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.441143 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.486172 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.486265 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.486301 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.486343 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486449 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486496 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486514 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486524 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486561 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486580 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486586 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:46.486566688 +0000 UTC m=+26.136834330 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486649 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:46.486627269 +0000 UTC m=+26.136894931 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486518 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486712 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:46.486700591 +0000 UTC m=+26.136968253 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486447 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.486821 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:46.486793544 +0000 UTC m=+26.137061206 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.544422 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.544467 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.544477 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.544497 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.544508 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.647658 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.647755 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.647773 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.647801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.647818 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.662165 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.662267 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.662282 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.662162 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.662665 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:42 crc kubenswrapper[4957]: E1206 05:39:42.662697 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.751071 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.751140 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.751162 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.751188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.751207 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.839010 4957 generic.go:334] "Generic (PLEG): container finished" podID="a8f0915e-a437-4104-8c8c-c7c10fbf8255" containerID="88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df" exitCode=0 Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.839099 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerDied","Data":"88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.841228 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.854138 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.854305 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.854352 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.854390 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.854432 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.863955 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.884645 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.898495 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.913668 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.926986 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.946913 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.957725 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.957798 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.957807 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.957824 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.957849 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:42Z","lastTransitionTime":"2025-12-06T05:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.973640 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:42 crc kubenswrapper[4957]: I1206 05:39:42.992738 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:42Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.006408 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.024949 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.039003 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.056177 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.061296 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.061342 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.061355 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.061374 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.061389 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.072249 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.091113 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.108609 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.126339 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.147999 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.164677 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.164728 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.164738 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.164754 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.164764 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.168825 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.196350 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.215961 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.233392 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.247762 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.266862 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.267770 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.267820 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.267857 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.267884 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.267897 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.284262 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.370710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.370779 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.370797 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.370823 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.370877 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.474563 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.474613 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.474626 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.474646 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.474660 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.577610 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.577689 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.577710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.577745 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.577774 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.680881 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.680943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.680960 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.680985 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.681001 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.709929 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.717814 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.723057 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.727918 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-jzcdr"] Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.728652 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.730811 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.731261 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.731803 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.731892 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.748079 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.766575 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.781863 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.784482 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.784545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.784559 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.784584 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.784601 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.798724 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.802101 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/88722903-d443-4b52-a7b9-815f5bfdf398-serviceca\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.802148 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmptb\" (UniqueName: \"kubernetes.io/projected/88722903-d443-4b52-a7b9-815f5bfdf398-kube-api-access-nmptb\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.802200 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88722903-d443-4b52-a7b9-815f5bfdf398-host\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.814385 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.834796 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.857904 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.862345 4957 generic.go:334] "Generic (PLEG): container finished" podID="a8f0915e-a437-4104-8c8c-c7c10fbf8255" containerID="77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4" exitCode=0 Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.862450 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerDied","Data":"77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.872707 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.887304 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.887375 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.887388 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.887429 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.887442 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.890015 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.902591 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.903424 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/88722903-d443-4b52-a7b9-815f5bfdf398-serviceca\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.903672 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmptb\" (UniqueName: \"kubernetes.io/projected/88722903-d443-4b52-a7b9-815f5bfdf398-kube-api-access-nmptb\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.903920 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88722903-d443-4b52-a7b9-815f5bfdf398-host\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.904137 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88722903-d443-4b52-a7b9-815f5bfdf398-host\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.904440 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/88722903-d443-4b52-a7b9-815f5bfdf398-serviceca\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.918523 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.927740 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmptb\" (UniqueName: \"kubernetes.io/projected/88722903-d443-4b52-a7b9-815f5bfdf398-kube-api-access-nmptb\") pod \"node-ca-jzcdr\" (UID: \"88722903-d443-4b52-a7b9-815f5bfdf398\") " pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.936783 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.954981 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.973138 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.990436 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.990477 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.990490 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.990503 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.990514 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:43Z","lastTransitionTime":"2025-12-06T05:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:43 crc kubenswrapper[4957]: I1206 05:39:43.991058 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.009681 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.028619 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.045931 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.051517 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jzcdr" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.066599 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: W1206 05:39:44.068353 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88722903_d443_4b52_a7b9_815f5bfdf398.slice/crio-91f8916516003c81cc3e2ee69d92a5c731134d1bc4398d79b1564ba8a841ad74 WatchSource:0}: Error finding container 91f8916516003c81cc3e2ee69d92a5c731134d1bc4398d79b1564ba8a841ad74: Status 404 returned error can't find the container with id 91f8916516003c81cc3e2ee69d92a5c731134d1bc4398d79b1564ba8a841ad74 Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.083603 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.093520 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.093582 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.093595 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.093620 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.093635 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.101761 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.116024 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.143932 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.164523 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.184751 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.196347 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.196400 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.196413 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.196434 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.196449 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.202300 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.216004 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.306073 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.306140 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.306157 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.306185 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.306209 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.409174 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.409238 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.409261 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.409295 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.409313 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.512749 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.512781 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.512799 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.512820 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.512860 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.616542 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.616602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.616616 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.616637 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.616654 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.661658 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.661659 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.661815 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:44 crc kubenswrapper[4957]: E1206 05:39:44.661994 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:44 crc kubenswrapper[4957]: E1206 05:39:44.662512 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:44 crc kubenswrapper[4957]: E1206 05:39:44.662713 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.723321 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.723368 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.723382 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.723400 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.723412 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.827191 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.827258 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.827275 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.827300 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.827317 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.872713 4957 generic.go:334] "Generic (PLEG): container finished" podID="a8f0915e-a437-4104-8c8c-c7c10fbf8255" containerID="aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329" exitCode=0 Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.873312 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerDied","Data":"aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.875344 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jzcdr" event={"ID":"88722903-d443-4b52-a7b9-815f5bfdf398","Type":"ContainerStarted","Data":"c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.875415 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jzcdr" event={"ID":"88722903-d443-4b52-a7b9-815f5bfdf398","Type":"ContainerStarted","Data":"91f8916516003c81cc3e2ee69d92a5c731134d1bc4398d79b1564ba8a841ad74"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.893403 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.915035 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.930089 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.930160 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.930180 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.930212 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.930232 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:44Z","lastTransitionTime":"2025-12-06T05:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.935520 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.952341 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:44 crc kubenswrapper[4957]: I1206 05:39:44.972943 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.000145 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:44Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.013709 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.029688 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.035400 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.035447 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.035462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.035488 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.035504 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.053139 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.069325 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.094497 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.110749 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.132695 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.137759 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.137794 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.137808 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.137832 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.137874 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.149608 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.168867 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.185519 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.201988 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.220402 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.237825 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.240728 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.240805 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.240816 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.240858 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.240874 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.256134 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.270001 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.287056 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.304026 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.330090 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.343553 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.343602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.343613 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.343630 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.343650 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.346720 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.361105 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.374176 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.388507 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.445731 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.445784 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.445801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.445879 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.445899 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.549041 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.549113 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.549134 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.549163 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.549181 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.656733 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.656805 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.656823 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.656898 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.656924 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.761474 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.761532 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.761551 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.761578 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.761599 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.864746 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.865143 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.865161 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.865183 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.865202 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.887320 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.891530 4957 generic.go:334] "Generic (PLEG): container finished" podID="a8f0915e-a437-4104-8c8c-c7c10fbf8255" containerID="8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9" exitCode=0 Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.891595 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerDied","Data":"8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.916166 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.945168 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.965198 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.967298 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.967349 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.967363 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.967384 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.967397 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:45Z","lastTransitionTime":"2025-12-06T05:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:45 crc kubenswrapper[4957]: I1206 05:39:45.982094 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.000335 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:45Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.021938 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.040815 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.060902 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.069648 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.069703 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.069719 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.069746 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.069765 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:46Z","lastTransitionTime":"2025-12-06T05:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.078364 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.113630 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.128299 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.142624 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.160301 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.173575 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.173601 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.173613 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.173631 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.173644 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:46Z","lastTransitionTime":"2025-12-06T05:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.176242 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.276882 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.276952 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.276970 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.276994 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.277013 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:46Z","lastTransitionTime":"2025-12-06T05:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.431247 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.431710 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:39:54.431678263 +0000 UTC m=+34.081945935 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.532398 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.532459 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.532497 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.532560 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.532691 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.532710 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.532746 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.532794 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:54.532779092 +0000 UTC m=+34.183046734 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533206 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533251 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:54.533240635 +0000 UTC m=+34.183508277 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533286 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533317 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:54.533307397 +0000 UTC m=+34.183575049 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533380 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533402 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533414 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.533439 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:39:54.533431411 +0000 UTC m=+34.183699053 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.614707 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.614774 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.614793 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.614925 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.614954 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:46Z","lastTransitionTime":"2025-12-06T05:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.661519 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.661562 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.661531 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.661760 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.661923 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:46 crc kubenswrapper[4957]: E1206 05:39:46.662056 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.717256 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.717950 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.717999 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.718027 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.718041 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:46Z","lastTransitionTime":"2025-12-06T05:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.820588 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.820644 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.820660 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.820681 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.820694 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:46Z","lastTransitionTime":"2025-12-06T05:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.896651 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.896731 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.922411 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.931203 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.931289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.931309 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.931336 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.931359 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:46Z","lastTransitionTime":"2025-12-06T05:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.947261 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.970078 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.982721 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:46 crc kubenswrapper[4957]: I1206 05:39:46.996566 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:46Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.010421 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.024006 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.034752 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.034790 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.034805 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.034826 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.034856 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.038496 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.040998 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.045187 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.057194 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.072908 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.091435 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.108126 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.123174 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.135039 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.137585 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.137637 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.137651 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.137674 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.137690 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.148866 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.167060 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.179237 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.194151 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.219154 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.241271 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.241329 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.241343 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.241367 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.241385 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.252504 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.272090 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.284458 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.298492 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.312157 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.328520 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.341973 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.344057 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.344098 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.344109 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.344125 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.344136 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.355177 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.366610 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.446936 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.446999 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.447011 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.447041 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.447054 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.550248 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.550337 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.550357 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.550382 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.550400 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.655592 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.656145 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.656341 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.656540 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.656731 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.760289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.760338 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.760357 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.760385 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.760403 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.863195 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.863284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.863311 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.863352 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.863390 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.915391 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" event={"ID":"a8f0915e-a437-4104-8c8c-c7c10fbf8255","Type":"ContainerStarted","Data":"e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.915936 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.938663 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.961505 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.966724 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.966932 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.967003 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.967071 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.967138 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:47Z","lastTransitionTime":"2025-12-06T05:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.977254 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:47 crc kubenswrapper[4957]: I1206 05:39:47.991924 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:47Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.009361 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.021576 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.036734 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.050924 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.070602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.070676 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.070701 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.070730 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.070755 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.076077 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.089902 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.102387 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.121863 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.145579 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.160970 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:48Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.184566 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.184647 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.184676 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.184698 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.184712 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.287762 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.288016 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.288077 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.288142 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.288217 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.391680 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.391712 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.391720 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.391733 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.391742 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.494747 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.494820 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.494890 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.494923 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.494941 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.598959 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.599027 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.599044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.599069 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.599084 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.661816 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.662109 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.661824 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:48 crc kubenswrapper[4957]: E1206 05:39:48.662240 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:48 crc kubenswrapper[4957]: E1206 05:39:48.662446 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:48 crc kubenswrapper[4957]: E1206 05:39:48.662675 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.702808 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.703000 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.703083 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.703122 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.703187 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.806745 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.806823 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.806884 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.806911 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.806929 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.909422 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.909489 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.909513 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.909545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.909569 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:48Z","lastTransitionTime":"2025-12-06T05:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:48 crc kubenswrapper[4957]: I1206 05:39:48.918536 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.013141 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.013209 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.013227 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.013254 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.013272 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.116311 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.116381 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.116402 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.116430 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.116451 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.224610 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.224661 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.224675 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.224705 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.224722 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.328606 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.328660 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.328674 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.328695 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.328707 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.431227 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.431268 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.431279 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.431293 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.431303 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.533912 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.534267 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.534280 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.534320 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.534335 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.637333 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.637408 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.637433 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.637467 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.637490 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.739738 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.739784 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.739796 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.739813 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.739825 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.842759 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.842797 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.842806 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.842820 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.842847 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.944992 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.945036 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.945050 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.945068 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:49 crc kubenswrapper[4957]: I1206 05:39:49.945084 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:49Z","lastTransitionTime":"2025-12-06T05:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.048907 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.048967 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.048984 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.049008 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.049025 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.152440 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.152503 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.152522 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.152548 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.152568 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.255494 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.255554 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.255572 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.255595 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.255612 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.358179 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.358243 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.358261 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.358284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.358300 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.461022 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.461067 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.461079 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.461098 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.461113 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.497544 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.521990 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.547602 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.563257 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.563315 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.563335 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.563359 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.563379 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.565755 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.587829 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.606549 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.635556 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.653916 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.661970 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:50 crc kubenswrapper[4957]: E1206 05:39:50.662148 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.662217 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.662307 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:50 crc kubenswrapper[4957]: E1206 05:39:50.662407 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:50 crc kubenswrapper[4957]: E1206 05:39:50.662533 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.666885 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.666926 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.666938 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.666955 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.666969 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.675161 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.694647 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.714618 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.736169 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.759942 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.769801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.769917 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.769938 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.769963 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.769981 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.775331 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.794577 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.813066 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.832632 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.846222 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.868498 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.874017 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.874059 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.874073 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.874095 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.874110 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.883228 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.896904 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.911260 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.928401 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/0.log" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.929149 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.932238 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354" exitCode=1 Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.932274 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.932955 4957 scope.go:117] "RemoveContainer" containerID="1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.943060 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.961993 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.977743 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.977789 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.977806 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.977826 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.977862 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:50Z","lastTransitionTime":"2025-12-06T05:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.979076 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:50 crc kubenswrapper[4957]: I1206 05:39:50.995690 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.011805 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.027189 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.045634 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.062890 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.073371 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.079774 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.079822 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.079860 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.079884 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.079900 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.089657 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.101397 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.121324 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:141\\\\nI1206 05:39:49.789211 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:39:49.789227 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 05:39:49.789226 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 05:39:49.789259 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:39:49.789273 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:39:49.789279 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:39:49.789357 6224 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 05:39:49.789667 6224 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:39:49.789890 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:39:49.789929 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:39:49.789972 6224 factory.go:656] Stopping watch factory\\\\nI1206 05:39:49.789995 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:39:49.790007 6224 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.134610 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.146975 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.161544 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.175329 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.182263 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.182304 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.182314 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.182331 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.182341 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.192248 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.212141 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.224807 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.235467 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.284920 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.284960 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.284970 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.284985 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.284994 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.388114 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.388161 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.388171 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.388187 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.388198 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.490326 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.490366 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.490379 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.490403 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.490415 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.592759 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.592800 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.592810 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.592846 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.592857 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.694897 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.694935 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.694946 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.694963 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.694975 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.758228 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.758377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.758465 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.758545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.758619 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: E1206 05:39:51.786044 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.791715 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.791749 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.791763 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.791784 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.791798 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: E1206 05:39:51.809895 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.820250 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.820276 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.820284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.820298 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.820310 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: E1206 05:39:51.832468 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.836597 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.836623 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.836632 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.836644 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.836652 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: E1206 05:39:51.851207 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.854855 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.854927 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.854937 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.854952 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.854963 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: E1206 05:39:51.867545 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: E1206 05:39:51.867663 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.869412 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.869442 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.869450 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.869463 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.869472 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.937567 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/0.log" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.940656 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.940856 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.957310 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.971808 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.971970 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.972191 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.972353 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.972419 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:51Z","lastTransitionTime":"2025-12-06T05:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.972977 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:51 crc kubenswrapper[4957]: I1206 05:39:51.990860 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:51Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.005350 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.024746 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.039202 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.048389 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.060511 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.072984 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.075807 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.075916 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.075932 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.075959 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.075977 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.106550 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:141\\\\nI1206 05:39:49.789211 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:39:49.789227 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 05:39:49.789226 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 05:39:49.789259 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:39:49.789273 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:39:49.789279 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:39:49.789357 6224 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 05:39:49.789667 6224 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:39:49.789890 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:39:49.789929 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:39:49.789972 6224 factory.go:656] Stopping watch factory\\\\nI1206 05:39:49.789995 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:39:49.790007 6224 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.119652 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.131788 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.147317 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.166738 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.178670 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.178708 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.178718 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.178732 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.178742 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.281063 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.281116 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.281132 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.281152 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.281167 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.384601 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.384664 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.384683 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.384710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.384729 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.488927 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.488992 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.489011 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.489036 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.489053 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.591795 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.591921 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.591957 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.591990 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.592012 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.662007 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.662097 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:52 crc kubenswrapper[4957]: E1206 05:39:52.662248 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:52 crc kubenswrapper[4957]: E1206 05:39:52.662426 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.662581 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:52 crc kubenswrapper[4957]: E1206 05:39:52.662718 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.694723 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.694768 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.694780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.694800 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.694812 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.797574 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.797869 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.797981 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.798114 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.798195 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.900518 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.900579 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.900597 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.900619 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.900631 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:52Z","lastTransitionTime":"2025-12-06T05:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.925287 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62"] Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.926036 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.928741 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.928751 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.958285 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:141\\\\nI1206 05:39:49.789211 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:39:49.789227 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 05:39:49.789226 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 05:39:49.789259 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:39:49.789273 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:39:49.789279 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:39:49.789357 6224 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 05:39:49.789667 6224 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:39:49.789890 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:39:49.789929 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:39:49.789972 6224 factory.go:656] Stopping watch factory\\\\nI1206 05:39:49.789995 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:39:49.790007 6224 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.972175 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:52 crc kubenswrapper[4957]: I1206 05:39:52.985005 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.001529 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:52Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.003386 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.003423 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.003437 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.003460 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.003476 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.006694 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ef8a2ba-dcfd-4538-b0c4-561056635820-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.006741 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ef8a2ba-dcfd-4538-b0c4-561056635820-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.006772 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74kxk\" (UniqueName: \"kubernetes.io/projected/4ef8a2ba-dcfd-4538-b0c4-561056635820-kube-api-access-74kxk\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.006944 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ef8a2ba-dcfd-4538-b0c4-561056635820-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.018962 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.037740 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.054520 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.071851 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.085322 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.100913 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.106186 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.106226 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.106235 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.106248 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.106259 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.107639 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ef8a2ba-dcfd-4538-b0c4-561056635820-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.107679 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ef8a2ba-dcfd-4538-b0c4-561056635820-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.107705 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74kxk\" (UniqueName: \"kubernetes.io/projected/4ef8a2ba-dcfd-4538-b0c4-561056635820-kube-api-access-74kxk\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.107772 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ef8a2ba-dcfd-4538-b0c4-561056635820-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.108381 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ef8a2ba-dcfd-4538-b0c4-561056635820-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.108613 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ef8a2ba-dcfd-4538-b0c4-561056635820-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.118375 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.118661 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ef8a2ba-dcfd-4538-b0c4-561056635820-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.135895 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74kxk\" (UniqueName: \"kubernetes.io/projected/4ef8a2ba-dcfd-4538-b0c4-561056635820-kube-api-access-74kxk\") pod \"ovnkube-control-plane-749d76644c-pxc62\" (UID: \"4ef8a2ba-dcfd-4538-b0c4-561056635820\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.136625 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.151544 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.170563 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.184497 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:53Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.209186 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.209282 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.209297 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.209320 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.209334 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.239910 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.312117 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.312184 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.312199 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.312220 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.312237 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.415397 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.415482 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.415496 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.415519 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.415532 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.518913 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.518990 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.519019 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.519053 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.519076 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.621906 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.621955 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.621967 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.621990 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.622003 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.724862 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.724915 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.724929 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.724955 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.724975 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.828452 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.828541 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.828579 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.828617 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.828642 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.931889 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.931958 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.931975 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.931998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:53 crc kubenswrapper[4957]: I1206 05:39:53.932015 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:53Z","lastTransitionTime":"2025-12-06T05:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.038403 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.038436 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.038447 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.038460 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.038469 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.047679 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-slqnh"] Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.048140 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.048207 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.060393 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.074016 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.089768 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.112626 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.119988 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.120035 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgdv9\" (UniqueName: \"kubernetes.io/projected/fcf7c4c9-894b-450c-83f6-441d3639f75a-kube-api-access-mgdv9\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.136999 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:141\\\\nI1206 05:39:49.789211 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:39:49.789227 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 05:39:49.789226 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 05:39:49.789259 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:39:49.789273 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:39:49.789279 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:39:49.789357 6224 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 05:39:49.789667 6224 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:39:49.789890 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:39:49.789929 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:39:49.789972 6224 factory.go:656] Stopping watch factory\\\\nI1206 05:39:49.789995 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:39:49.790007 6224 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.140908 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.140998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.141014 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.141032 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.141045 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.152928 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.165445 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.176860 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.192628 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.212001 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.220605 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.220646 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgdv9\" (UniqueName: \"kubernetes.io/projected/fcf7c4c9-894b-450c-83f6-441d3639f75a-kube-api-access-mgdv9\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.220893 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.221034 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:39:54.72100149 +0000 UTC m=+34.371269162 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.224031 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.244351 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.244390 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.244401 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.244416 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.244426 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.247454 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.256682 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgdv9\" (UniqueName: \"kubernetes.io/projected/fcf7c4c9-894b-450c-83f6-441d3639f75a-kube-api-access-mgdv9\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.268287 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.287552 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.306489 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.326468 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.347329 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.347369 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.347386 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.347409 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.347424 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: W1206 05:39:54.370984 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ef8a2ba_dcfd_4538_b0c4_561056635820.slice/crio-c40e5d23b8f898bf6723da61e9169842e6fe7aaa1ff9d749520a0baf86142eb1 WatchSource:0}: Error finding container c40e5d23b8f898bf6723da61e9169842e6fe7aaa1ff9d749520a0baf86142eb1: Status 404 returned error can't find the container with id c40e5d23b8f898bf6723da61e9169842e6fe7aaa1ff9d749520a0baf86142eb1 Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.451487 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.451553 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.451570 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.451598 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.451616 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.523770 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.523939 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:40:10.523914187 +0000 UTC m=+50.174181819 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.553888 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.553948 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.553963 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.553987 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.554004 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.624723 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.624817 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.624872 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.624897 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625072 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625095 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625089 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625126 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625184 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625208 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625212 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:10.625184591 +0000 UTC m=+50.275452213 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625109 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625147 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625345 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:10.625290704 +0000 UTC m=+50.275558516 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625370 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:10.625360996 +0000 UTC m=+50.275628838 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.625424 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:10.625414047 +0000 UTC m=+50.275681909 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.656320 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.656364 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.656387 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.656405 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.656416 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.662216 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.662225 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.662366 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.662395 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.662631 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.662551 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.725903 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.726056 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.726145 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:39:55.726125646 +0000 UTC m=+35.376393278 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.759432 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.759475 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.759487 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.759503 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.759516 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.861789 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.861855 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.861866 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.861879 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.861888 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.953416 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" event={"ID":"4ef8a2ba-dcfd-4538-b0c4-561056635820","Type":"ContainerStarted","Data":"14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.953472 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" event={"ID":"4ef8a2ba-dcfd-4538-b0c4-561056635820","Type":"ContainerStarted","Data":"ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.953485 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" event={"ID":"4ef8a2ba-dcfd-4538-b0c4-561056635820","Type":"ContainerStarted","Data":"c40e5d23b8f898bf6723da61e9169842e6fe7aaa1ff9d749520a0baf86142eb1"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.955117 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/1.log" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.955812 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/0.log" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.960697 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6" exitCode=1 Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.960817 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.961000 4957 scope.go:117] "RemoveContainer" containerID="1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.962354 4957 scope.go:117] "RemoveContainer" containerID="0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6" Dec 06 05:39:54 crc kubenswrapper[4957]: E1206 05:39:54.962582 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.964032 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.964066 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.964076 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.964090 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.964133 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:54Z","lastTransitionTime":"2025-12-06T05:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.972892 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:54 crc kubenswrapper[4957]: I1206 05:39:54.989942 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.001186 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:54Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.012998 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.025698 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.040762 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.053781 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.066602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.066654 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.066665 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.066681 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.066691 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.072061 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:141\\\\nI1206 05:39:49.789211 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:39:49.789227 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 05:39:49.789226 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 05:39:49.789259 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:39:49.789273 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:39:49.789279 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:39:49.789357 6224 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 05:39:49.789667 6224 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:39:49.789890 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:39:49.789929 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:39:49.789972 6224 factory.go:656] Stopping watch factory\\\\nI1206 05:39:49.789995 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:39:49.790007 6224 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.086696 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.103475 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.131039 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.161278 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.169221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.169281 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.169293 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.169318 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.169332 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.184603 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.198406 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.211223 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.226108 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.240879 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.254431 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.269564 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.272548 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.272594 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.272605 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.272624 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.272637 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.286416 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.301601 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.315968 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.331468 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.344280 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.358524 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.370765 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.375667 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.375732 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.375748 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.375766 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.375776 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.388423 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.407605 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.426436 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dc6255ecf0c7656330151dad315190543564b967ef5eef1f5f68ec5ce381354\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:141\\\\nI1206 05:39:49.789211 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:39:49.789227 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1206 05:39:49.789226 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1206 05:39:49.789259 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:39:49.789273 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:39:49.789279 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:39:49.789357 6224 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1206 05:39:49.789667 6224 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:39:49.789890 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:39:49.789929 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:39:49.789972 6224 factory.go:656] Stopping watch factory\\\\nI1206 05:39:49.789995 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:39:49.790007 6224 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"ault]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:39:53.131363 6390 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 05:39:53.131415 6390 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:39:53.131481 6390 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 05:39:53.131580 6390 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.438519 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.449370 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.461419 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:55Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.478358 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.478405 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.478416 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.478432 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.478442 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.580641 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.580729 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.580748 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.580777 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.580792 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.661382 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:55 crc kubenswrapper[4957]: E1206 05:39:55.661677 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.684732 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.684780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.684792 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.684811 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.684825 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.736702 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:55 crc kubenswrapper[4957]: E1206 05:39:55.736966 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:55 crc kubenswrapper[4957]: E1206 05:39:55.737045 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:39:57.737024317 +0000 UTC m=+37.387291959 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.787629 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.787719 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.787743 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.787774 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.787803 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.891411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.891498 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.891522 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.891552 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.891575 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.968146 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/1.log" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.994193 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.994242 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.994253 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.994271 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:55 crc kubenswrapper[4957]: I1206 05:39:55.994285 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:55Z","lastTransitionTime":"2025-12-06T05:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.097668 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.097800 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.097827 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.097900 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.097925 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.201701 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.201774 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.201790 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.201813 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.201864 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.305562 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.305641 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.305658 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.305682 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.305702 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.408480 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.408557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.408581 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.408610 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.408635 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.511370 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.511457 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.511475 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.511499 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.511518 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.614164 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.614217 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.614237 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.614258 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.614271 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.662277 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.662374 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:56 crc kubenswrapper[4957]: E1206 05:39:56.662486 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:56 crc kubenswrapper[4957]: E1206 05:39:56.662616 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.662761 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:56 crc kubenswrapper[4957]: E1206 05:39:56.662892 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.718109 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.718189 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.718211 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.718240 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.718263 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.821361 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.821446 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.821472 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.821501 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.821522 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.924621 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.924679 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.924698 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.924721 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:56 crc kubenswrapper[4957]: I1206 05:39:56.924739 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:56Z","lastTransitionTime":"2025-12-06T05:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.027766 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.027874 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.027901 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.027932 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.027960 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.131085 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.131158 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.131176 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.131202 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.131219 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.233909 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.233988 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.234005 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.234023 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.234034 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.337209 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.337268 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.337284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.337311 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.337329 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.440502 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.440590 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.440624 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.440657 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.440679 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.544397 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.544458 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.544486 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.544514 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.544537 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.647686 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.647767 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.647789 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.647820 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.647881 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.662121 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:57 crc kubenswrapper[4957]: E1206 05:39:57.662289 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.751761 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.751882 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.751910 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.751943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.751965 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.761569 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:57 crc kubenswrapper[4957]: E1206 05:39:57.761881 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:57 crc kubenswrapper[4957]: E1206 05:39:57.761990 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:40:01.761960854 +0000 UTC m=+41.412228526 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.855315 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.855357 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.855369 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.855388 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.855401 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.957797 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.957936 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.957966 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.957998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:57 crc kubenswrapper[4957]: I1206 05:39:57.958066 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:57Z","lastTransitionTime":"2025-12-06T05:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.060774 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.060860 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.060878 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.060903 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.060920 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.164311 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.164355 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.164371 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.164394 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.164415 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.267283 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.267355 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.267370 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.267391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.267407 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.370097 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.370168 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.370182 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.370201 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.370216 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.462722 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.464182 4957 scope.go:117] "RemoveContainer" containerID="0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6" Dec 06 05:39:58 crc kubenswrapper[4957]: E1206 05:39:58.464449 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.472868 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.472949 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.472972 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.473001 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.473023 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.500061 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"ault]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:39:53.131363 6390 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 05:39:53.131415 6390 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:39:53.131481 6390 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 05:39:53.131580 6390 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.520093 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.535675 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.553363 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.571491 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.575182 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.575237 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.575254 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.575273 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.575288 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.590935 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.610086 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.625644 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.641744 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.656165 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.662032 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:39:58 crc kubenswrapper[4957]: E1206 05:39:58.662752 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.662802 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:39:58 crc kubenswrapper[4957]: E1206 05:39:58.663141 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.663285 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:39:58 crc kubenswrapper[4957]: E1206 05:39:58.663481 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.673903 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.677174 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.677326 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.677428 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.677523 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.677608 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.691550 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.712198 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.731461 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.754138 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.768074 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:39:58Z is after 2025-08-24T17:21:41Z" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.779995 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.780044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.780077 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.780096 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.780110 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.883173 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.883228 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.883240 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.883256 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.883271 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.985380 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.985424 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.985437 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.985454 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:58 crc kubenswrapper[4957]: I1206 05:39:58.985466 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:58Z","lastTransitionTime":"2025-12-06T05:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.088311 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.088354 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.088366 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.088382 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.088393 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.191917 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.192038 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.192057 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.192077 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.192092 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.294577 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.294643 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.294654 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.294694 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.294709 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.397490 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.397540 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.397553 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.397567 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.397577 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.500664 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.500703 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.500714 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.500730 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.500740 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.603924 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.604004 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.604014 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.604030 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.604043 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.662070 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:39:59 crc kubenswrapper[4957]: E1206 05:39:59.662235 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.706427 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.706453 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.706461 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.706473 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.706481 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.809457 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.809777 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.809975 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.810125 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.810259 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.913867 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.913936 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.913957 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.913986 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:39:59 crc kubenswrapper[4957]: I1206 05:39:59.914007 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:39:59Z","lastTransitionTime":"2025-12-06T05:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.017005 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.017080 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.017098 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.017123 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.017144 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.120479 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.120558 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.120575 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.120600 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.120616 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.223471 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.223523 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.223535 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.223554 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.223566 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.326579 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.326639 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.326659 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.326684 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.326704 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.431111 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.431181 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.431206 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.431237 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.431260 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.533467 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.533541 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.533558 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.533584 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.533603 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.635864 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.635902 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.635915 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.635931 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.635942 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.661246 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.661364 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.661645 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:00 crc kubenswrapper[4957]: E1206 05:40:00.661658 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:00 crc kubenswrapper[4957]: E1206 05:40:00.661722 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:00 crc kubenswrapper[4957]: E1206 05:40:00.661787 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.685050 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.706543 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.723596 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.737852 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.737885 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.737897 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.737913 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.737926 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.743282 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.765967 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.781886 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.798658 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.817519 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.838481 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.840415 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.840462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.840509 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.840525 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.840537 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.852341 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.865884 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.881452 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.894528 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.912185 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.924957 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.942411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.942475 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.942525 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.942553 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.942571 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:00Z","lastTransitionTime":"2025-12-06T05:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:00 crc kubenswrapper[4957]: I1206 05:40:00.951373 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"ault]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:39:53.131363 6390 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 05:39:53.131415 6390 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:39:53.131481 6390 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 05:39:53.131580 6390 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:00Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.045337 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.045393 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.045403 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.045415 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.045424 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.148684 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.148880 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.148911 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.148942 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.148966 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.251502 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.251557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.251576 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.251600 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.251622 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.353784 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.353864 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.353878 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.353896 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.353909 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.456701 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.456780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.456803 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.456867 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.456894 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.560030 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.560080 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.560090 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.560107 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.560119 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.661754 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:01 crc kubenswrapper[4957]: E1206 05:40:01.662134 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.665353 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.665442 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.665468 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.665504 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.665537 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.768775 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.769186 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.769328 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.769450 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.769573 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.810680 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:01 crc kubenswrapper[4957]: E1206 05:40:01.810947 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:01 crc kubenswrapper[4957]: E1206 05:40:01.811044 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:40:09.811011427 +0000 UTC m=+49.461279099 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.873403 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.873732 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.873908 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.874093 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.874231 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.977696 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.977739 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.977748 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.977764 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:01 crc kubenswrapper[4957]: I1206 05:40:01.977774 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:01Z","lastTransitionTime":"2025-12-06T05:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.004078 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.004104 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.004112 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.004124 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.004133 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.026432 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:02Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.032013 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.032078 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.032090 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.032108 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.032126 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.049007 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:02Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.053452 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.053532 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.053558 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.053586 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.053606 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.072559 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:02Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.078124 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.078209 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.078240 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.078271 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.078290 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.098691 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:02Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.104442 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.104484 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.104495 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.104509 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.104519 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.121318 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:02Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.121487 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.123813 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.124017 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.124128 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.124222 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.124309 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.226821 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.227103 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.227172 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.227239 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.227309 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.329785 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.329918 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.329936 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.329964 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.329990 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.432437 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.432497 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.432516 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.432542 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.432560 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.535505 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.535559 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.535576 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.535599 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.535615 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.638161 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.638212 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.638223 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.638242 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.638253 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.661879 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.661899 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.662036 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.662157 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.662467 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:02 crc kubenswrapper[4957]: E1206 05:40:02.662632 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.740945 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.741024 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.741044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.741069 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.741091 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.844070 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.844139 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.844158 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.844207 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.844232 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.947539 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.947602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.947618 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.947645 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:02 crc kubenswrapper[4957]: I1206 05:40:02.947665 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:02Z","lastTransitionTime":"2025-12-06T05:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.051411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.051472 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.051489 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.051512 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.051531 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.155106 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.155231 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.155264 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.155305 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.155328 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.258481 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.258544 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.258562 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.258586 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.258604 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.361360 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.361466 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.361485 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.361508 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.361526 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.464429 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.464504 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.464542 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.464576 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.464604 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.567374 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.567415 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.567425 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.567441 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.567451 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.661914 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:03 crc kubenswrapper[4957]: E1206 05:40:03.662119 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.670088 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.670148 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.670163 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.670191 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.670207 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.774301 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.774375 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.774399 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.774432 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.774457 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.896340 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.896379 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.896389 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.896405 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.896416 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.998887 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.998928 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.998937 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.998952 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:03 crc kubenswrapper[4957]: I1206 05:40:03.998962 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:03Z","lastTransitionTime":"2025-12-06T05:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.102333 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.102391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.102411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.102433 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.102450 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.205877 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.205929 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.205945 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.205974 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.205992 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.309667 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.309714 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.309730 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.309752 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.309771 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.413116 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.413171 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.413185 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.413206 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.413223 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.516175 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.516698 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.516775 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.516953 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.517026 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.619594 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.619670 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.619683 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.619700 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.619712 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.661400 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.661411 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:04 crc kubenswrapper[4957]: E1206 05:40:04.661639 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.661700 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:04 crc kubenswrapper[4957]: E1206 05:40:04.661809 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:04 crc kubenswrapper[4957]: E1206 05:40:04.661931 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.723635 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.723712 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.723739 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.723768 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.723792 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.827734 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.827801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.827818 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.827871 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.827888 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.930780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.931081 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.931149 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.931213 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:04 crc kubenswrapper[4957]: I1206 05:40:04.931273 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:04Z","lastTransitionTime":"2025-12-06T05:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.034372 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.034669 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.034769 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.034895 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.034985 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.138047 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.138110 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.138133 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.138166 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.138185 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.241123 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.241469 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.241610 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.241775 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.241986 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.344485 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.344525 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.344537 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.344557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.344571 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.447628 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.447710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.447732 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.447764 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.447785 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.551111 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.551163 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.551174 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.551192 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.551204 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.653403 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.653445 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.653454 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.653467 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.653476 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.662141 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:05 crc kubenswrapper[4957]: E1206 05:40:05.662245 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.755614 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.755663 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.755675 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.755691 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.755703 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.858740 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.858821 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.858898 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.858930 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.858952 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.962287 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.962648 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.962780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.962981 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:05 crc kubenswrapper[4957]: I1206 05:40:05.963119 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:05Z","lastTransitionTime":"2025-12-06T05:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.066767 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.066862 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.066881 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.066910 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.066927 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.170220 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.170769 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.171023 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.171390 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.171544 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.275343 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.275412 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.275435 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.275463 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.275485 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.378645 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.378690 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.378702 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.378722 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.378735 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.482312 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.482695 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.482926 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.483133 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.483283 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.586960 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.587019 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.587035 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.587064 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.587082 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.661401 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.661550 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:06 crc kubenswrapper[4957]: E1206 05:40:06.661628 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.661407 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:06 crc kubenswrapper[4957]: E1206 05:40:06.661751 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:06 crc kubenswrapper[4957]: E1206 05:40:06.662044 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.690355 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.690406 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.690422 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.690443 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.690459 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.793035 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.793080 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.793092 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.793108 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.793121 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.895471 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.895518 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.895530 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.895546 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.895559 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.998254 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.998325 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.998336 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.998353 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:06 crc kubenswrapper[4957]: I1206 05:40:06.998364 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:06Z","lastTransitionTime":"2025-12-06T05:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.101287 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.101342 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.101358 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.101383 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.101401 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.204066 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.204116 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.204132 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.204153 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.204165 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.306943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.307018 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.307041 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.307077 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.307100 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.410655 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.410721 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.410738 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.410765 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.410783 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.513368 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.513490 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.513517 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.513552 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.513579 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.616522 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.616585 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.616788 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.616818 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.616882 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.662231 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:07 crc kubenswrapper[4957]: E1206 05:40:07.662495 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.720353 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.720411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.720429 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.720455 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.720472 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.824367 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.824433 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.824561 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.824591 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.824608 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.928228 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.928304 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.928328 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.928357 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:07 crc kubenswrapper[4957]: I1206 05:40:07.928381 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:07Z","lastTransitionTime":"2025-12-06T05:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.031079 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.031143 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.031160 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.031188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.031207 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.138619 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.138702 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.138716 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.138735 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.138747 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.241727 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.241810 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.241876 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.241920 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.241941 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.345194 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.345242 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.345257 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.345281 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.345299 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.448317 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.448377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.448400 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.448429 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.448450 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.551386 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.551434 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.551453 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.551476 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.551494 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.654339 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.655131 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.655188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.655216 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.655235 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.661768 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:08 crc kubenswrapper[4957]: E1206 05:40:08.661934 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.662236 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.662256 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:08 crc kubenswrapper[4957]: E1206 05:40:08.662678 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:08 crc kubenswrapper[4957]: E1206 05:40:08.662829 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.758293 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.758715 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.758891 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.759064 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.759218 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.862011 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.862063 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.862079 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.862095 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.862107 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.964909 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.964996 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.965016 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.965048 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:08 crc kubenswrapper[4957]: I1206 05:40:08.965068 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:08Z","lastTransitionTime":"2025-12-06T05:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.068767 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.068888 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.068909 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.068939 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.068959 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.173018 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.173085 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.173123 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.173160 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.173188 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.276615 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.276700 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.276721 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.276752 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.276773 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.379899 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.379943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.379956 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.379978 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.379991 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.483497 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.483568 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.483591 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.483619 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.483640 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.587012 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.587065 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.587077 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.587095 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.587107 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.661674 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:09 crc kubenswrapper[4957]: E1206 05:40:09.661913 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.690406 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.690453 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.690465 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.690488 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.690510 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.793346 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.793418 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.793435 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.793455 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.793467 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.896566 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.896661 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.896680 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.896709 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.896727 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.910105 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:09 crc kubenswrapper[4957]: E1206 05:40:09.910294 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:09 crc kubenswrapper[4957]: E1206 05:40:09.910361 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:40:25.910340477 +0000 UTC m=+65.560608109 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.999107 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.999150 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.999160 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.999180 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:09 crc kubenswrapper[4957]: I1206 05:40:09.999194 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:09Z","lastTransitionTime":"2025-12-06T05:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.102655 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.103315 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.103381 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.103463 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.103523 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.206688 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.207041 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.207115 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.207209 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.207362 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.309816 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.309939 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.309959 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.309989 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.310015 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.413246 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.413666 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.413945 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.414153 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.414290 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.522527 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.522605 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.522623 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.522646 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.522667 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.617336 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.617585 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:40:42.61754851 +0000 UTC m=+82.267816182 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.625373 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.625429 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.625451 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.625476 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.625495 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.661343 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.661380 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.661495 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.663068 4957 scope.go:117] "RemoveContainer" containerID="0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.663122 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.663409 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.663701 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.680618 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.698954 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.719007 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.719120 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.719180 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719259 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.719285 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719301 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719334 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719379 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719411 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719423 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719434 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719360 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719500 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:42.719465511 +0000 UTC m=+82.369733213 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719541 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:42.719522923 +0000 UTC m=+82.369790595 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719605 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:42.719555944 +0000 UTC m=+82.369823616 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: E1206 05:40:10.719647 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:40:42.719629435 +0000 UTC m=+82.369897107 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.724474 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"ault]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:39:53.131363 6390 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 05:39:53.131415 6390 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:39:53.131481 6390 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 05:39:53.131580 6390 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.730358 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.730423 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.730444 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.730471 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.730491 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.741674 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.754972 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.771213 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.787713 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.805191 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.816727 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.829889 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.832524 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.832555 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.832566 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.832582 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.832595 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.842635 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.853944 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.867298 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.881076 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.898124 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.913665 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:10Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.935202 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.935250 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.935260 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.935281 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:10 crc kubenswrapper[4957]: I1206 05:40:10.935293 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:10Z","lastTransitionTime":"2025-12-06T05:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.038481 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.038521 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.038533 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.038551 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.038565 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.039599 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/1.log" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.047610 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.048320 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.065234 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.085194 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.110907 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.129102 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.141054 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.141098 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.141112 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.141130 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.141141 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.160730 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"ault]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:39:53.131363 6390 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 05:39:53.131415 6390 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:39:53.131481 6390 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 05:39:53.131580 6390 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.185876 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.198078 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.210982 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.224227 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.237671 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.243701 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.243844 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.243923 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.244015 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.244109 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.252194 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.268564 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.282438 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.295409 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.315074 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.330890 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.346807 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.346962 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.347044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.347144 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.347233 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.451993 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.452240 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.452522 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.452716 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.452942 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.555617 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.555874 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.555985 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.556073 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.556139 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.658833 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.658896 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.658911 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.658930 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.658943 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.662147 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:11 crc kubenswrapper[4957]: E1206 05:40:11.662273 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.761347 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.761431 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.761459 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.761486 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.761504 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.863682 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.863742 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.863764 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.863793 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.863817 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.906615 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.924709 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.929213 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.949430 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.967209 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.967303 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.967321 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.967349 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.967366 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:11Z","lastTransitionTime":"2025-12-06T05:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.972393 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:11 crc kubenswrapper[4957]: I1206 05:40:11.989746 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:11Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.008278 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.030066 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.048605 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.052305 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/2.log" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.053174 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/1.log" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.057816 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123" exitCode=1 Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.057938 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.058104 4957 scope.go:117] "RemoveContainer" containerID="0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.060374 4957 scope.go:117] "RemoveContainer" containerID="0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123" Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.064156 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.070256 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.070312 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.070328 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.070374 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.070392 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.080233 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.098686 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.110007 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.141281 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"ault]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:39:53.131363 6390 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 05:39:53.131415 6390 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:39:53.131481 6390 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 05:39:53.131580 6390 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.153296 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.162155 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.172687 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.172796 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.172817 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.172888 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.172908 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.174770 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.191503 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.209319 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.245383 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.265007 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.278120 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.278154 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.278164 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.278179 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.278189 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.284141 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.300992 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.316814 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.332807 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.349558 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.363610 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.377161 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.381305 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.381399 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.381418 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.381484 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.381509 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.395278 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.406133 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.421717 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.438928 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.466340 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab4db0986bf5089abaf5d24646899ea04e65c67bc635458a196568596ec5fa6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"ault]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_UDP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[udp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {4c1be812-05d3-4f45-91b5-a853a5c8de71}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 05:39:53.131363 6390 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 05:39:53.131415 6390 ovnkube.go:599] Stopped ovnkube\\\\nI1206 05:39:53.131481 6390 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 05:39:53.131580 6390 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.474366 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.474412 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.474431 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.474453 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.474465 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.481450 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.489305 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.494102 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.494192 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.494221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.494232 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.494245 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.494253 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.509331 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.512156 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.514497 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.514532 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.514543 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.514560 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.514572 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.532881 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.536964 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.537003 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.537013 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.537028 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.537040 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.548485 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.553071 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.553187 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.553227 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.553254 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.553274 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.570818 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:12Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.571032 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.572724 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.572777 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.572787 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.572801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.572810 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.661755 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.661860 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.661894 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.662040 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.661751 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:12 crc kubenswrapper[4957]: E1206 05:40:12.662169 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.675382 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.675671 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.676012 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.676325 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.676596 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.779022 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.779614 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.779736 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.779967 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.780100 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.883355 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.883751 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.883996 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.884196 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.884376 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.987979 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.988323 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.988483 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.988603 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:12 crc kubenswrapper[4957]: I1206 05:40:12.988688 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:12Z","lastTransitionTime":"2025-12-06T05:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.063575 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/2.log" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.068827 4957 scope.go:117] "RemoveContainer" containerID="0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123" Dec 06 05:40:13 crc kubenswrapper[4957]: E1206 05:40:13.069618 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.091040 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.091087 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.091106 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.091136 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.091157 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.092294 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.120666 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.140937 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.161352 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.184820 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.194104 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.194152 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.194168 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.194188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.194203 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.204534 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.238782 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.258214 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.275903 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.296852 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.296903 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.296914 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.296935 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.296949 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.301880 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.323039 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.343079 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.359038 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.377953 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.392268 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.400370 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.400415 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.400426 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.400445 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.400457 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.406912 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.420929 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:13Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.503397 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.503461 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.503478 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.503505 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.503522 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.606754 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.606810 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.606821 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.606860 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.606873 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.661951 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:13 crc kubenswrapper[4957]: E1206 05:40:13.662130 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.710424 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.710471 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.710483 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.710499 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.710511 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.812974 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.813026 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.813036 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.813053 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.813065 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.915414 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.915461 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.915477 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.915531 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:13 crc kubenswrapper[4957]: I1206 05:40:13.915546 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:13Z","lastTransitionTime":"2025-12-06T05:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.018173 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.018218 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.018227 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.018242 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.018254 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.121204 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.121244 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.121255 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.121271 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.121280 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.225087 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.225407 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.225550 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.225651 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.225753 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.328589 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.328630 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.328641 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.328658 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.328667 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.431099 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.431142 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.431150 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.431165 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.431178 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.533385 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.533432 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.533444 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.533460 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.533471 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.635767 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.635821 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.635857 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.635875 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.635886 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.661353 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:14 crc kubenswrapper[4957]: E1206 05:40:14.661509 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.661546 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.661666 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:14 crc kubenswrapper[4957]: E1206 05:40:14.661693 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:14 crc kubenswrapper[4957]: E1206 05:40:14.661947 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.738798 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.738862 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.738874 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.738887 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.738896 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.841329 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.841386 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.841401 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.841418 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.841432 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.944943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.944998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.945014 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.945037 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:14 crc kubenswrapper[4957]: I1206 05:40:14.945055 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:14Z","lastTransitionTime":"2025-12-06T05:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.048108 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.048175 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.048194 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.048221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.048240 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.151177 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.151245 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.151270 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.151302 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.151324 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.254817 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.254950 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.254999 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.255026 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.255045 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.358733 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.358789 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.358797 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.358813 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.358823 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.461293 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.461364 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.461385 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.461411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.461434 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.563224 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.563284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.563301 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.563324 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.563343 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.662181 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:15 crc kubenswrapper[4957]: E1206 05:40:15.662367 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.666411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.666443 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.666459 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.666480 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.666496 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.768355 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.768420 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.768437 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.768462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.768482 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.872451 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.872518 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.872536 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.872561 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.872580 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.976621 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.976674 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.976688 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.976721 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:15 crc kubenswrapper[4957]: I1206 05:40:15.976735 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:15Z","lastTransitionTime":"2025-12-06T05:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.079572 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.079654 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.079679 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.079712 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.079736 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.182535 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.182601 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.182618 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.182643 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.182661 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.285409 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.285468 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.285480 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.285502 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.285516 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.388675 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.388753 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.388779 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.388907 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.388987 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.491811 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.491987 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.492014 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.492058 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.492084 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.595219 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.595289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.595308 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.595334 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.595374 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.661665 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.661744 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.661811 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:16 crc kubenswrapper[4957]: E1206 05:40:16.661998 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:16 crc kubenswrapper[4957]: E1206 05:40:16.662271 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:16 crc kubenswrapper[4957]: E1206 05:40:16.662257 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.698990 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.699052 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.699067 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.699090 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.699107 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.801705 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.801761 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.801777 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.801800 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.801817 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.905405 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.905462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.905482 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.905509 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:16 crc kubenswrapper[4957]: I1206 05:40:16.905525 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:16Z","lastTransitionTime":"2025-12-06T05:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.008491 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.008557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.008576 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.008602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.008621 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.111450 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.111533 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.111553 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.111579 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.111599 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.215294 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.215364 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.215380 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.215406 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.215427 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.318955 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.319007 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.319018 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.319039 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.319052 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.422044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.422098 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.422110 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.422131 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.422144 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.523637 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.523916 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.523999 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.524161 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.524244 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.627539 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.627891 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.628004 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.628177 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.628330 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.661932 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:17 crc kubenswrapper[4957]: E1206 05:40:17.662192 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.730966 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.731008 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.731021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.731036 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.731048 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.833935 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.834009 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.834033 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.834064 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.834087 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.936941 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.936985 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.936994 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.937009 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:17 crc kubenswrapper[4957]: I1206 05:40:17.937018 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:17Z","lastTransitionTime":"2025-12-06T05:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.039781 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.039822 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.039860 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.039878 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.039889 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.142686 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.142725 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.142737 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.142755 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.142795 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.245254 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.245284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.245292 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.245306 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.245316 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.347951 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.348011 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.348024 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.348042 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.348054 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.464977 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.465029 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.465044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.465064 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.465076 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.567766 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.567813 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.567825 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.567864 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.567877 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.661675 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.661860 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:18 crc kubenswrapper[4957]: E1206 05:40:18.661910 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.661936 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:18 crc kubenswrapper[4957]: E1206 05:40:18.662066 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:18 crc kubenswrapper[4957]: E1206 05:40:18.662168 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.670400 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.670461 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.670478 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.670497 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.670510 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.773069 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.773123 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.773134 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.773152 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.773166 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.875976 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.876021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.876032 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.876047 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.876060 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.978319 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.978365 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.978377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.978394 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:18 crc kubenswrapper[4957]: I1206 05:40:18.978407 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:18Z","lastTransitionTime":"2025-12-06T05:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.081314 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.081359 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.081370 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.081390 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.081401 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.199940 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.200016 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.200033 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.200061 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.200078 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.302943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.303306 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.303514 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.303707 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.303934 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.411889 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.412134 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.412198 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.412260 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.412316 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.515554 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.516072 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.516473 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.516823 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.517074 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.619649 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.619694 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.619705 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.619723 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.619734 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.661467 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:19 crc kubenswrapper[4957]: E1206 05:40:19.661920 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.721882 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.721916 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.721927 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.721942 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.721956 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.824462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.824493 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.824503 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.824516 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.824526 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.926992 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.927334 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.927528 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.927692 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:19 crc kubenswrapper[4957]: I1206 05:40:19.927878 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:19Z","lastTransitionTime":"2025-12-06T05:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.030484 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.030546 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.030562 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.030584 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.030599 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.134253 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.134346 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.134366 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.134391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.134408 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.236885 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.237195 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.237300 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.237387 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.237489 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.339713 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.339746 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.339754 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.339766 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.339774 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.442362 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.442437 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.442448 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.442471 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.442483 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.545637 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.545968 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.546038 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.546279 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.546351 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.648913 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.648998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.649023 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.649054 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.649077 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.661256 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:20 crc kubenswrapper[4957]: E1206 05:40:20.661381 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.661471 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.661601 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:20 crc kubenswrapper[4957]: E1206 05:40:20.661804 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:20 crc kubenswrapper[4957]: E1206 05:40:20.662024 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.688898 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.711265 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.726159 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.743657 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.751148 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.751183 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.751194 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.751210 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.751223 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.758332 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.775748 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.792856 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.806338 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.820993 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.843152 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.853417 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.853477 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.853495 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.853543 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.853561 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.856420 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.870701 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.883939 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.909970 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.924902 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.949567 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.955557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.955598 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.955606 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.955620 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.955628 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:20Z","lastTransitionTime":"2025-12-06T05:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:20 crc kubenswrapper[4957]: I1206 05:40:20.967372 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:20Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.058064 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.058092 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.058103 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.058120 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.058131 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.161438 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.161549 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.161570 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.161593 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.161610 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.265027 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.265130 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.265164 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.265194 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.265216 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.368158 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.368269 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.368289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.368312 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.368328 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.470605 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.470638 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.470649 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.470663 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.470674 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.573294 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.573325 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.573333 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.573345 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.573354 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.661348 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:21 crc kubenswrapper[4957]: E1206 05:40:21.661591 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.675682 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.675893 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.675921 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.675944 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.675961 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.778818 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.779298 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.779549 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.779807 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.780053 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.883091 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.883589 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.883762 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.883944 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.884109 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.987121 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.987176 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.987196 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.987221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:21 crc kubenswrapper[4957]: I1206 05:40:21.987243 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:21Z","lastTransitionTime":"2025-12-06T05:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.091102 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.091146 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.091195 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.091216 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.091230 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.194511 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.194575 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.194593 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.194621 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.194642 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.298599 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.299262 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.299443 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.299678 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.300385 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.403077 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.403127 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.403140 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.403158 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.403173 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.506677 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.506764 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.506782 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.506813 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.506856 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.610314 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.610384 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.610407 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.610439 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.610462 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.661576 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.661672 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.661800 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.662021 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.662161 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.662727 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.714527 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.714595 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.714614 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.714638 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.714657 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.756253 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.756290 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.756300 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.756316 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.756327 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.772345 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.779786 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.779822 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.779847 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.779863 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.779878 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.798636 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.803348 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.803411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.803431 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.803455 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.803472 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.820744 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.825419 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.825503 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.825550 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.825575 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.825590 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.844647 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.849371 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.849628 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.849766 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.849944 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.850166 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.869332 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:22Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:22 crc kubenswrapper[4957]: E1206 05:40:22.870082 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.872347 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.872394 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.872409 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.872431 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.872446 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.975281 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.975330 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.975350 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.975374 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:22 crc kubenswrapper[4957]: I1206 05:40:22.975394 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:22Z","lastTransitionTime":"2025-12-06T05:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.077986 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.078039 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.078057 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.078079 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.078097 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.181129 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.181193 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.181215 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.181243 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.181266 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.285247 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.285313 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.285327 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.285352 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.285368 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.389772 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.389821 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.389853 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.389875 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.389890 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.492762 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.492857 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.492881 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.492908 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.492929 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.595724 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.596147 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.596221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.596294 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.596376 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.662356 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:23 crc kubenswrapper[4957]: E1206 05:40:23.662650 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.699247 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.699336 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.699363 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.699399 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.699427 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.802955 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.803392 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.803491 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.803578 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.803672 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.906806 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.906920 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.906947 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.906979 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:23 crc kubenswrapper[4957]: I1206 05:40:23.907002 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:23Z","lastTransitionTime":"2025-12-06T05:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.009993 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.010386 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.010552 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.010693 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.010821 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.113191 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.113232 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.113246 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.113261 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.113271 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.215167 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.215411 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.215423 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.215439 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.215454 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.318416 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.318469 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.318480 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.318494 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.318506 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.421115 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.421148 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.421158 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.421172 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.421181 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.524253 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.524292 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.524301 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.524313 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.524322 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.627188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.627235 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.627249 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.627268 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.627279 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.661963 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.661984 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.662051 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:24 crc kubenswrapper[4957]: E1206 05:40:24.662170 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:24 crc kubenswrapper[4957]: E1206 05:40:24.662576 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:24 crc kubenswrapper[4957]: E1206 05:40:24.662767 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.663324 4957 scope.go:117] "RemoveContainer" containerID="0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123" Dec 06 05:40:24 crc kubenswrapper[4957]: E1206 05:40:24.663767 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.729158 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.729200 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.729213 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.729230 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.729240 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.831590 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.831627 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.831635 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.831648 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.831658 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.933857 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.933912 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.933922 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.933938 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:24 crc kubenswrapper[4957]: I1206 05:40:24.933956 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:24Z","lastTransitionTime":"2025-12-06T05:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.035973 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.036168 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.036368 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.036557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.036737 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.138858 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.138892 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.138900 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.138912 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.138922 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.247213 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.247249 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.247258 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.247271 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.247282 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.350658 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.350707 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.350719 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.350737 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.350748 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.453024 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.453068 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.453077 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.453090 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.453099 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.556178 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.556229 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.556241 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.556258 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.556273 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.658646 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.658686 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.658698 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.658711 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.658720 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.661943 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:25 crc kubenswrapper[4957]: E1206 05:40:25.662050 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.761540 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.761594 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.761606 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.761624 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.761638 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.863647 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.863683 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.863693 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.863706 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.863715 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.966201 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.966246 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.966255 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.966271 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.966290 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:25Z","lastTransitionTime":"2025-12-06T05:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:25 crc kubenswrapper[4957]: I1206 05:40:25.985685 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:25 crc kubenswrapper[4957]: E1206 05:40:25.985852 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:25 crc kubenswrapper[4957]: E1206 05:40:25.985909 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:40:57.985894185 +0000 UTC m=+97.636161817 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.067921 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.067951 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.067959 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.067970 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.067979 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.111011 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/0.log" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.111069 4957 generic.go:334] "Generic (PLEG): container finished" podID="1384ce42-5e9a-4d6a-a299-f564d1494b0a" containerID="0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b" exitCode=1 Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.111102 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerDied","Data":"0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.111478 4957 scope.go:117] "RemoveContainer" containerID="0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.123611 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.135639 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.150427 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.164102 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.172219 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.172265 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.172277 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.172294 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.172308 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.192928 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.206586 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.219902 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.230751 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.245662 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.258709 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:25Z\\\",\\\"message\\\":\\\"2025-12-06T05:39:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865\\\\n2025-12-06T05:39:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865 to /host/opt/cni/bin/\\\\n2025-12-06T05:39:40Z [verbose] multus-daemon started\\\\n2025-12-06T05:39:40Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:40:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.274270 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.274315 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.274325 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.274341 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.274351 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.278909 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.295046 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.311104 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.324100 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.335889 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.346275 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.357208 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:26Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.376673 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.376710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.376722 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.376737 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.376748 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.479171 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.479208 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.479219 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.479234 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.479244 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.581978 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.582042 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.582057 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.582078 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.582091 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.662092 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.662190 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:26 crc kubenswrapper[4957]: E1206 05:40:26.662274 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:26 crc kubenswrapper[4957]: E1206 05:40:26.662368 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.662470 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:26 crc kubenswrapper[4957]: E1206 05:40:26.662549 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.685178 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.685231 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.685240 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.685257 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.685268 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.788344 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.788386 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.788398 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.788416 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.788429 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.890557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.890587 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.890595 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.890610 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.890618 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.993455 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.993482 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.993493 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.993507 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:26 crc kubenswrapper[4957]: I1206 05:40:26.993518 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:26Z","lastTransitionTime":"2025-12-06T05:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.096208 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.096276 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.096289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.096305 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.096315 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.117375 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/0.log" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.117463 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerStarted","Data":"9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.143514 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.159947 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.173213 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.186691 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.198940 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.198996 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.199014 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.199037 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.199059 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.205385 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.220700 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.239226 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.257491 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.273629 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:25Z\\\",\\\"message\\\":\\\"2025-12-06T05:39:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865\\\\n2025-12-06T05:39:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865 to /host/opt/cni/bin/\\\\n2025-12-06T05:39:40Z [verbose] multus-daemon started\\\\n2025-12-06T05:39:40Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:40:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.289860 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.302231 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.302276 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.302293 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.302314 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.302331 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.311312 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.323403 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.339188 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.359577 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.377133 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.395316 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.407053 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.407090 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.407101 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.407120 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.407132 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.407826 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:27Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.510053 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.510096 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.510110 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.510141 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.510160 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.613300 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.613355 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.613392 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.613415 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.613434 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.661724 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:27 crc kubenswrapper[4957]: E1206 05:40:27.661954 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.715534 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.715599 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.715623 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.715648 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.715664 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.824345 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.824420 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.824443 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.824470 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.824491 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.927441 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.927490 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.927507 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.927529 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:27 crc kubenswrapper[4957]: I1206 05:40:27.927546 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:27Z","lastTransitionTime":"2025-12-06T05:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.030380 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.030457 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.030481 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.030509 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.030530 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.133110 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.133148 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.133185 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.133202 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.133213 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.235493 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.235525 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.235536 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.235549 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.235559 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.338188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.338225 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.338283 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.338299 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.338338 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.440657 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.440698 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.440712 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.440727 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.440739 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.542559 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.542596 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.542609 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.542654 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.542669 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.645876 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.645921 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.645932 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.645951 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.645962 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.662198 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.662247 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:28 crc kubenswrapper[4957]: E1206 05:40:28.662290 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.662309 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:28 crc kubenswrapper[4957]: E1206 05:40:28.662406 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:28 crc kubenswrapper[4957]: E1206 05:40:28.662526 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.748480 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.748522 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.748537 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.748558 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.748574 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.851456 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.851492 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.851501 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.851516 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.851526 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.954033 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.954417 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.954611 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.954812 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:28 crc kubenswrapper[4957]: I1206 05:40:28.955044 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:28Z","lastTransitionTime":"2025-12-06T05:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.057716 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.057799 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.057827 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.057902 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.057923 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.161452 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.161509 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.161528 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.161555 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.161579 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.264671 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.264734 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.264749 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.264768 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.264781 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.368448 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.368503 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.368517 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.368536 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.368549 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.472086 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.472133 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.472141 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.472156 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.472167 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.575019 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.575113 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.575134 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.575159 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.575181 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.661715 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:29 crc kubenswrapper[4957]: E1206 05:40:29.661986 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.677913 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.677995 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.678021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.678046 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.678063 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.780668 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.780739 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.780757 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.780780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.780796 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.883363 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.883447 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.883462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.883477 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.883503 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.986713 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.986780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.986791 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.986806 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:29 crc kubenswrapper[4957]: I1206 05:40:29.986816 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:29Z","lastTransitionTime":"2025-12-06T05:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.090250 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.090369 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.090410 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.090442 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.090464 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.193260 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.193335 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.193351 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.193376 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.193393 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.296642 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.296676 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.296684 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.296704 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.296715 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.400528 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.400614 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.400635 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.400669 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.400688 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.504051 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.504151 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.504170 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.504200 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.504222 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.607538 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.607621 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.607634 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.607653 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.607672 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.662351 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:30 crc kubenswrapper[4957]: E1206 05:40:30.662528 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.662933 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:30 crc kubenswrapper[4957]: E1206 05:40:30.663057 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.663213 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:30 crc kubenswrapper[4957]: E1206 05:40:30.663533 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.682431 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.711063 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.711134 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.711151 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.711177 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.711194 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.721458 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.739735 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.757100 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.772090 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.789813 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.814498 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.814545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.814556 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.814579 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.814592 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.817754 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:25Z\\\",\\\"message\\\":\\\"2025-12-06T05:39:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865\\\\n2025-12-06T05:39:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865 to /host/opt/cni/bin/\\\\n2025-12-06T05:39:40Z [verbose] multus-daemon started\\\\n2025-12-06T05:39:40Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:40:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.832754 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.852209 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.868795 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.886791 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.905947 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.917483 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.917545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.917565 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.917591 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.917610 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:30Z","lastTransitionTime":"2025-12-06T05:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.922521 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.938817 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.953094 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.970064 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:30 crc kubenswrapper[4957]: I1206 05:40:30.991734 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:30Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.020279 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.020525 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.020872 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.021121 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.021352 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.124344 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.124380 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.124391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.124407 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.124418 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.226347 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.226385 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.226400 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.226415 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.226425 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.329466 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.329602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.329628 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.329662 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.329684 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.432377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.433035 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.433145 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.433324 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.433463 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.536633 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.536684 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.536694 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.536714 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.536726 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.639084 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.639165 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.639185 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.639210 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.639229 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.661305 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:31 crc kubenswrapper[4957]: E1206 05:40:31.661422 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.742316 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.742576 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.742722 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.742907 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.743054 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.846496 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.846531 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.846540 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.846555 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.846564 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.950353 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.950788 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.951033 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.951410 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:31 crc kubenswrapper[4957]: I1206 05:40:31.951736 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:31Z","lastTransitionTime":"2025-12-06T05:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.054348 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.054391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.054403 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.054420 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.054431 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.157462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.157523 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.157540 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.157565 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.157582 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.260485 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.260560 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.260583 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.260613 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.260636 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.363545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.364069 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.364272 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.364446 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.364618 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.467760 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.467875 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.467892 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.467925 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.467938 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.571385 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.571455 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.571470 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.571490 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.571506 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.662282 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.662333 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.662449 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:32 crc kubenswrapper[4957]: E1206 05:40:32.662656 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:32 crc kubenswrapper[4957]: E1206 05:40:32.663101 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:32 crc kubenswrapper[4957]: E1206 05:40:32.663222 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.673394 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.673449 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.673471 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.673502 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.673525 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.776400 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.776481 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.776502 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.776533 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.776554 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.878948 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.879007 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.879023 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.879083 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.879103 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.981369 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.981409 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.981417 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.981433 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:32 crc kubenswrapper[4957]: I1206 05:40:32.981446 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:32Z","lastTransitionTime":"2025-12-06T05:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.084179 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.084211 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.084220 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.084234 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.084245 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.132204 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.132548 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.132568 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.132584 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.132595 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: E1206 05:40:33.152584 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.157616 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.157670 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.157687 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.157710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.157726 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: E1206 05:40:33.179872 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.183515 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.183542 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.183554 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.183567 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.183579 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: E1206 05:40:33.197122 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.200949 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.201004 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.201028 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.201058 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.201082 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: E1206 05:40:33.215650 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.220504 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.220621 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.220684 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.220750 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.220994 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: E1206 05:40:33.232262 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:33Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:33 crc kubenswrapper[4957]: E1206 05:40:33.232370 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.234221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.234347 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.234414 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.234481 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.234543 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.336746 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.336801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.336818 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.336881 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.336899 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.439578 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.439615 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.439631 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.439650 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.439665 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.547929 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.547994 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.548015 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.548042 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.548064 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.651546 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.651607 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.651631 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.651659 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.651680 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.661921 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:33 crc kubenswrapper[4957]: E1206 05:40:33.662092 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.755363 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.755431 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.755445 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.755467 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.755483 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.858954 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.859002 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.859016 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.859034 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.859045 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.962602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.962660 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.962680 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.962705 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:33 crc kubenswrapper[4957]: I1206 05:40:33.962723 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:33Z","lastTransitionTime":"2025-12-06T05:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.066986 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.067050 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.067066 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.067092 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.067108 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.170659 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.170717 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.170728 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.170748 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.170760 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.274583 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.274632 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.274654 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.274682 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.274763 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.379902 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.379959 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.379979 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.380003 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.380024 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.482932 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.482982 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.482998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.483021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.483036 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.586951 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.587021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.587045 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.588007 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.588038 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.661619 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.661678 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.661861 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:34 crc kubenswrapper[4957]: E1206 05:40:34.662028 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:34 crc kubenswrapper[4957]: E1206 05:40:34.662149 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:34 crc kubenswrapper[4957]: E1206 05:40:34.662265 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.678250 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.690491 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.690591 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.690616 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.690827 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.690884 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.794213 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.794280 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.794303 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.794331 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.794354 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.898789 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.898922 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.898944 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.898967 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:34 crc kubenswrapper[4957]: I1206 05:40:34.898985 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:34Z","lastTransitionTime":"2025-12-06T05:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.002494 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.002613 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.002630 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.002647 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.002659 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.105768 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.105829 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.105888 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.105910 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.105927 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.209070 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.209363 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.209446 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.209548 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.209630 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.313114 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.313247 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.313358 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.313467 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.313549 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.416512 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.417060 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.417327 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.417550 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.417791 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.521055 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.521125 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.521149 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.521172 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.521185 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.624099 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.624161 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.624179 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.624205 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.624224 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.661679 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:35 crc kubenswrapper[4957]: E1206 05:40:35.662024 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.727975 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.728050 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.728069 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.728104 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.728139 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.831239 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.831328 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.831354 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.831387 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.831449 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.933896 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.933931 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.933941 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.933955 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:35 crc kubenswrapper[4957]: I1206 05:40:35.933966 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:35Z","lastTransitionTime":"2025-12-06T05:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.036296 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.036349 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.036361 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.036383 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.036394 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.144655 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.144719 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.144739 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.144766 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.144784 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.247156 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.247208 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.247222 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.247240 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.247252 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.350223 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.350563 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.350698 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.350788 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.350893 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.453614 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.453687 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.453710 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.453739 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.453761 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.556229 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.556292 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.556310 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.556333 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.556352 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.659318 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.659367 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.659384 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.659407 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.659423 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.661459 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:36 crc kubenswrapper[4957]: E1206 05:40:36.661655 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.661481 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.661694 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:36 crc kubenswrapper[4957]: E1206 05:40:36.661933 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:36 crc kubenswrapper[4957]: E1206 05:40:36.661954 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.762866 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.763005 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.763032 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.763059 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.763078 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.866242 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.866300 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.866321 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.866345 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.866363 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.970022 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.970082 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.970100 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.970127 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:36 crc kubenswrapper[4957]: I1206 05:40:36.970146 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:36Z","lastTransitionTime":"2025-12-06T05:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.073588 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.073986 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.074174 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.074323 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.074452 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.176911 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.176971 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.176982 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.176999 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.177013 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.280226 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.280348 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.280373 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.280399 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.280414 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.383812 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.384251 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.384407 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.384555 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.384690 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.488410 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.488470 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.488491 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.488519 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.488539 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.591910 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.591978 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.592002 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.592032 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.592053 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.661524 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:37 crc kubenswrapper[4957]: E1206 05:40:37.661721 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.662655 4957 scope.go:117] "RemoveContainer" containerID="0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.695135 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.695419 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.695686 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.695961 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.696181 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.799984 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.800454 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.800473 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.800498 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.800521 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.904173 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.904252 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.904270 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.904298 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:37 crc kubenswrapper[4957]: I1206 05:40:37.904317 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:37Z","lastTransitionTime":"2025-12-06T05:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.007303 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.007368 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.007391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.007421 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.007444 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.110193 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.110247 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.110263 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.110284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.110299 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.213271 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.213332 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.213349 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.213377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.213396 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.316503 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.316581 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.316606 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.316634 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.316657 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.419973 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.420028 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.420044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.420069 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.420092 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.523447 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.523507 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.523527 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.523557 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.523577 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.626345 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.626390 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.626402 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.626420 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.626434 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.661644 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:38 crc kubenswrapper[4957]: E1206 05:40:38.662610 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.661810 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:38 crc kubenswrapper[4957]: E1206 05:40:38.662801 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.661719 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:38 crc kubenswrapper[4957]: E1206 05:40:38.663052 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.728192 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.728247 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.728266 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.728289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.728308 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.831135 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.831198 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.831216 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.831240 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.831259 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.934486 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.934552 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.934571 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.934594 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:38 crc kubenswrapper[4957]: I1206 05:40:38.934612 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:38Z","lastTransitionTime":"2025-12-06T05:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.037617 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.037659 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.037667 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.037679 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.037688 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.141311 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.141381 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.141399 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.141424 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.141441 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.245034 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.245140 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.245158 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.245184 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.245201 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.352249 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.352607 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.352631 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.352651 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.352667 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.455488 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.455549 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.455566 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.455591 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.455610 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.559213 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.559269 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.559287 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.559313 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.559330 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.661317 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:39 crc kubenswrapper[4957]: E1206 05:40:39.661605 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.664018 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.664100 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.664118 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.664145 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.664162 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.768672 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.768740 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.768761 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.768792 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.768814 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.872545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.872652 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.872667 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.872694 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.872709 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.976973 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.977042 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.977062 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.977091 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:39 crc kubenswrapper[4957]: I1206 05:40:39.977109 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:39Z","lastTransitionTime":"2025-12-06T05:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.080141 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.080206 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.080224 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.080270 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.080286 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.183574 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.183632 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.183746 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.183777 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.183866 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.287092 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.287154 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.287177 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.287208 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.287231 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.389822 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.389894 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.389909 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.389929 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.389944 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.492699 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.492745 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.492763 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.492787 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.492805 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.596243 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.596301 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.596320 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.596344 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.596362 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.661515 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.661511 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:40 crc kubenswrapper[4957]: E1206 05:40:40.661740 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.661781 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:40 crc kubenswrapper[4957]: E1206 05:40:40.662443 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:40 crc kubenswrapper[4957]: E1206 05:40:40.662924 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.689583 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.699302 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.699379 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.699401 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.699426 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.699444 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.711725 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.733439 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.761244 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.786313 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.803596 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.803670 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.803691 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.803720 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.803743 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.805389 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.829064 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.851376 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.868532 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.893598 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9019dc7d-622a-4652-86a4-18e5f86931a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9af17c49e0732a21c5ccc770ca8367336ae1b9c225bafebb180a403a5d002d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.905870 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.905905 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.905916 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.905930 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.905942 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:40Z","lastTransitionTime":"2025-12-06T05:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.912518 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.936561 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.952348 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.971216 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.982784 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:40 crc kubenswrapper[4957]: I1206 05:40:40.994984 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:40Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.007732 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.007775 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.007784 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.007801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.007820 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.007995 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:25Z\\\",\\\"message\\\":\\\"2025-12-06T05:39:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865\\\\n2025-12-06T05:39:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865 to /host/opt/cni/bin/\\\\n2025-12-06T05:39:40Z [verbose] multus-daemon started\\\\n2025-12-06T05:39:40Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:40:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.020910 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.110818 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.110927 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.110970 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.110998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.111014 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.173178 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/2.log" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.176795 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.177222 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.195592 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.213560 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.213605 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.213618 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.213641 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.213657 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.220738 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.236062 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.253397 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.269971 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9019dc7d-622a-4652-86a4-18e5f86931a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9af17c49e0732a21c5ccc770ca8367336ae1b9c225bafebb180a403a5d002d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.286451 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.316433 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.316499 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.316516 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.316540 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.316558 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.318859 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.334321 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.351913 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.365404 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.380921 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.403515 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:25Z\\\",\\\"message\\\":\\\"2025-12-06T05:39:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865\\\\n2025-12-06T05:39:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865 to /host/opt/cni/bin/\\\\n2025-12-06T05:39:40Z [verbose] multus-daemon started\\\\n2025-12-06T05:39:40Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:40:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.419163 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.419241 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.419260 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.419285 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.419303 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.421928 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.453320 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.470625 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.488751 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.506234 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.522069 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.522109 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.522118 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.522132 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.522141 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.526511 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:41Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.624692 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.625087 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.625229 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.625352 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.625488 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.661406 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:41 crc kubenswrapper[4957]: E1206 05:40:41.661603 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.728980 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.730239 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.730287 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.730317 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.730336 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.834030 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.834134 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.834153 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.834179 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.834198 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.937450 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.937519 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.937537 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.937560 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:41 crc kubenswrapper[4957]: I1206 05:40:41.937580 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:41Z","lastTransitionTime":"2025-12-06T05:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.040804 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.040943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.040970 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.041001 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.041026 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.144129 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.144199 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.144222 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.144250 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.144273 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.246390 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.246661 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.246725 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.246799 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.246894 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.349546 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.349614 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.349631 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.349657 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.349707 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.453339 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.453407 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.453425 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.453451 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.453469 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.556398 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.556462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.556474 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.556492 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.556509 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.659395 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.659459 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.659477 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.659501 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.659519 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.661780 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.661859 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.661948 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.661964 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.662105 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.662242 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.680402 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.680545 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.68051558 +0000 UTC m=+146.330783252 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.762804 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.763021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.763045 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.763068 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.763087 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.781081 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.781191 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.781246 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.781280 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781463 4957 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781463 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781507 4957 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781537 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781550 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781598 4957 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781626 4957 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781561 4957 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781543 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.781521417 +0000 UTC m=+146.431789079 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781792 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.781764183 +0000 UTC m=+146.432031845 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781816 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.781802814 +0000 UTC m=+146.432070476 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:42 crc kubenswrapper[4957]: E1206 05:40:42.781871 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.781826035 +0000 UTC m=+146.432093707 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.866573 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.866621 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.866640 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.866663 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.866680 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.969669 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.969718 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.969737 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.969761 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:42 crc kubenswrapper[4957]: I1206 05:40:42.969777 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:42Z","lastTransitionTime":"2025-12-06T05:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.073239 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.073374 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.073396 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.073424 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.073486 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.176629 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.176670 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.176685 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.176705 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.176717 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.262297 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/3.log" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.263058 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/2.log" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.266740 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16" exitCode=1 Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.266794 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.266861 4957 scope.go:117] "RemoveContainer" containerID="0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.267677 4957 scope.go:117] "RemoveContainer" containerID="4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16" Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.267912 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.281174 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.281249 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.281272 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.281303 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.281327 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.288160 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"69fab357-117a-4747-b0d4-d44a08196da0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://755e5fe38e0dfd7d2e5a4e13792c9b2b0d8740c584c83ba676475f774236e247\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c88dfc0d7ff8b856c1caaf0791bfdf450e9c8d8d5997efb1c1510da7c9911459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e43eb06d1228006f34e5c73c835e49db76141ee7e37deb66e130e9b68359f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87a3c9a9ea9440f2ede41c11ebe9b610bd632e72a56dcca3975887e9ce772f0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.304510 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.320727 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3cffd9c69c52721b909468cfe926259a25427379b6e174873490e9b6ca1c939\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48f4ecd8faef2f67930e6ee89ea14d9b60b1323a2e9408edd585bab312bf8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.332309 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50f5b0413f4c06c258feffb99dcca3c3241929b28ce01984383b538b4e40b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.347172 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.360912 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84ba0710-bb5e-44cb-b8e0-b5a0a7a21f63\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b14a2ac694c72a1b8f6bd7a787fb9664b309d2281bf229bff64223d07baa4e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37be38316b3d5aed0ab3e3d9ade26fbfc897ab606e46ee7640b810a3b3886841\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b0f776726f816805bcb7c892336838de2292e0a938e0872d929d6642b03a78\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.380487 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0915e-a437-4104-8c8c-c7c10fbf8255\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4abda7570f7e509e2b5af2010fba4f3b892b93f17428ece3546a7ab1e5b1887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ad03a53391558037448014572de77a5050f99a8e6b1ced9584797c864396cb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3451386a381a5c8652680cb096489001ea494af2b042c897c4e839872deaab3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88e3a32b6de8ff6d05ac8eb7bb4678c96eb956af1269c28abb736976689be3df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77baba5cd5230012c75fc9a09755a49435e370bca9d9f7f153cc44333ae1e2b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aac1fb6bd8e4349a271a2afac1c99ae40e0081080061fb2c5727e5d05740c329\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8068fa0dce0bc7612bd444c02d454b27813792d91ef266eefa8e9e2896c628c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5km2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-g9hn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.384879 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.384930 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.384944 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.384967 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.384979 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.397168 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jzcdr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88722903-d443-4b52-a7b9-815f5bfdf398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c95fd87f455a6b378cc6b500088612afc745fe2ce378677971c81f16cfe59389\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmptb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jzcdr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.410575 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.424303 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.439872 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9019dc7d-622a-4652-86a4-18e5f86931a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9af17c49e0732a21c5ccc770ca8367336ae1b9c225bafebb180a403a5d002d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.454251 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.472650 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.472928 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.473011 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.473081 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.473143 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.476435 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa635ecb-5324-449e-a8f3-7a9bfdca7064\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a161da7e5695f764c88169a07c60f8e8b216c1784d07cb8c4d4ab4eac4d2123\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:11Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.615981 6626 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1206 05:40:11.615992 6626 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:11.616020 6626 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1206 05:40:11.616050 6626 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1206 05:40:11.616399 6626 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 05:40:11.616416 6626 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 05:40:11.616446 6626 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:11.616459 6626 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:11.616465 6626 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:11.616479 6626 factory.go:656] Stopping watch factory\\\\nI1206 05:40:11.616491 6626 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 05:40:11.616500 6626 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 05:40:11.616506 6626 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:11.616519 6626 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:11.616523 6626 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:42Z\\\",\\\"message\\\":\\\"topping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:41.395302 6945 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:41.395259 6945 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 05:40:41.395442 6945 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:40:41.395517 6945 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 05:40:41.395578 6945 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 05:40:41.395666 6945 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 05:40:41.395695 6945 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 05:40:41.395378 6945 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1206 05:40:41.395789 6945 factory.go:656] Stopping watch factory\\\\nI1206 05:40:41.395900 6945 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 05:40:41.395987 6945 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bpjx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hlm8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.484004 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.487305 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.487354 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.487365 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.487377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.487387 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.488460 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446a0886-35d8-4574-9ccd-6b38f6df37bc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f296fbfa5d22684cadea4e48c5eb08af224d66c791a9ef7d4faaf51ee695044e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nnxw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f2z46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.498012 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-78cpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d0bce5-6056-4247-8eef-23ed810df082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4dcfa0d1705febd5cc390e63524b9ffd5167439a02112bbf466f480fd5ad8f77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4skp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-78cpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.500314 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.503264 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.503305 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.503318 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.503335 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.503348 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.510659 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39599d5d08ba82f2db3e7cc3b65002c1f3f01e45ac85a9f1b8fbcc8cb142a105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.516067 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.519131 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.519171 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.519184 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.519203 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.519216 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.522998 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-69r8l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1384ce42-5e9a-4d6a-a299-f564d1494b0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T05:40:25Z\\\",\\\"message\\\":\\\"2025-12-06T05:39:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865\\\\n2025-12-06T05:39:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d4cac739-7d29-4e3b-b606-de9b5126d865 to /host/opt/cni/bin/\\\\n2025-12-06T05:39:40Z [verbose] multus-daemon started\\\\n2025-12-06T05:39:40Z [verbose] Readiness Indicator file check\\\\n2025-12-06T05:40:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:40:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dlzxt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-69r8l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.534192 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ef8a2ba-dcfd-4538-b0c4-561056635820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ede363e692b1e104ebd5a7b38f2434c451df1bb68fb2212dba6e43f99285c400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14437c3432e392faa563174e347ad78f3496edefc0eb4ffd3cb10b1bcd2a4a60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74kxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pxc62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.553510 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.562193 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.562246 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.562257 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.562275 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.562288 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.581789 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T05:40:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5e1c3eb-dd1f-4a16-8ee4-ad4718a743a8\\\",\\\"systemUUID\\\":\\\"15ddb378-9b6e-4ca1-8e52-89f797d2e2f0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:43Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.581974 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.583872 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.583906 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.583939 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.583957 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.583969 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.661544 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:43 crc kubenswrapper[4957]: E1206 05:40:43.661691 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.687160 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.687205 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.687225 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.687250 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.687268 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.789946 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.790736 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.790789 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.790824 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.790983 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.894253 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.894315 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.894334 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.894362 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.894381 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.996494 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.996533 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.996542 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.996558 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:43 crc kubenswrapper[4957]: I1206 05:40:43.996569 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:43Z","lastTransitionTime":"2025-12-06T05:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.099545 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.099608 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.099629 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.099655 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.099672 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.204573 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.204682 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.204714 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.204751 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.204772 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.273999 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/3.log" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.308079 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.308402 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.308538 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.308731 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.308912 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.412702 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.412774 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.412801 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.412877 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.412900 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.516928 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.516973 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.516990 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.517013 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.517030 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.619715 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.619761 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.619778 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.619799 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.619815 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.661524 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:44 crc kubenswrapper[4957]: E1206 05:40:44.661703 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.661538 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.661800 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:44 crc kubenswrapper[4957]: E1206 05:40:44.661882 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:44 crc kubenswrapper[4957]: E1206 05:40:44.662044 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.722907 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.722989 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.723019 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.723050 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.723073 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.827187 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.827726 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.827738 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.827760 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.827770 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.931190 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.931246 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.931258 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.931276 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:44 crc kubenswrapper[4957]: I1206 05:40:44.931288 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:44Z","lastTransitionTime":"2025-12-06T05:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.035066 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.035137 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.035157 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.035187 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.035208 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.139183 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.139243 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.139259 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.139284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.139302 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.243221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.243284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.243307 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.243330 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.243350 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.346166 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.346625 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.346914 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.347267 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.347409 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.450446 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.450491 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.450499 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.450511 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.450521 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.554059 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.554113 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.554124 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.554142 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.554153 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.657259 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.657330 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.657348 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.657370 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.657387 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.661661 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:45 crc kubenswrapper[4957]: E1206 05:40:45.661893 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.759622 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.759684 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.759701 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.759726 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.759744 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.863335 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.863412 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.863435 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.863464 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.863486 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.966403 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.966468 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.966491 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.966523 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:45 crc kubenswrapper[4957]: I1206 05:40:45.966546 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:45Z","lastTransitionTime":"2025-12-06T05:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.069690 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.069810 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.069828 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.069888 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.069904 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.216322 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.216389 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.216410 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.216433 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.216454 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.320051 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.320116 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.320145 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.320184 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.320206 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.423532 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.423625 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.423645 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.423668 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.423684 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.527448 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.527519 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.527543 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.527572 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.527594 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.631778 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.631863 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.631879 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.631906 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.631923 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.663101 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.663384 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.663478 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:46 crc kubenswrapper[4957]: E1206 05:40:46.663925 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:46 crc kubenswrapper[4957]: E1206 05:40:46.666400 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:46 crc kubenswrapper[4957]: E1206 05:40:46.666630 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.735005 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.735127 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.735164 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.735196 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.735218 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.837930 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.837981 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.837998 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.838023 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.838039 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.941055 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.941126 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.941146 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.941171 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:46 crc kubenswrapper[4957]: I1206 05:40:46.941189 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:46Z","lastTransitionTime":"2025-12-06T05:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.043729 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.043775 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.043786 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.043800 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.043810 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.147235 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.147290 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.147307 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.147331 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.147350 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.250871 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.250941 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.250960 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.250991 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.251015 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.354542 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.354602 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.354621 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.354645 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.354663 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.457334 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.457413 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.457436 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.457462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.457482 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.560889 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.561027 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.561125 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.561223 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.561245 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.662059 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:47 crc kubenswrapper[4957]: E1206 05:40:47.662591 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.664342 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.664521 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.664541 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.664564 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.664581 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.767869 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.767931 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.767950 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.767974 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.767991 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.871380 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.871432 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.871445 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.871463 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.871476 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.975062 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.975170 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.975189 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.975243 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:47 crc kubenswrapper[4957]: I1206 05:40:47.975266 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:47Z","lastTransitionTime":"2025-12-06T05:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.078377 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.078415 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.078426 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.078441 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.078454 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.182814 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.182935 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.182958 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.182984 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.183004 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.286337 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.286414 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.286430 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.286455 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.286471 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.389284 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.389353 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.389367 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.389390 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.389405 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.491869 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.491901 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.491910 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.491927 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.491937 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.595220 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.595272 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.595289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.595317 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.595340 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.662111 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:48 crc kubenswrapper[4957]: E1206 05:40:48.662363 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.662715 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:48 crc kubenswrapper[4957]: E1206 05:40:48.662898 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.663027 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:48 crc kubenswrapper[4957]: E1206 05:40:48.663247 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.698869 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.698944 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.698966 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.698996 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.699056 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.802502 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.802538 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.802553 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.802571 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.802583 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.904942 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.904973 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.904982 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.904994 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:48 crc kubenswrapper[4957]: I1206 05:40:48.905005 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:48Z","lastTransitionTime":"2025-12-06T05:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.008582 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.008640 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.008656 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.008677 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.008694 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.112235 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.112327 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.112354 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.112388 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.112410 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.215441 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.215520 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.215535 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.215555 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.215569 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.320189 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.320362 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.320391 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.320430 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.320466 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.425612 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.425689 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.425708 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.425738 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.425758 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.529057 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.529124 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.529141 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.529168 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.529186 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.632611 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.632732 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.632751 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.632780 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.632799 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.661309 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:49 crc kubenswrapper[4957]: E1206 05:40:49.661503 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.737098 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.737172 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.737188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.737212 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.737233 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.840410 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.840511 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.840525 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.840542 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.840556 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.943984 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.944058 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.944080 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.944106 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:49 crc kubenswrapper[4957]: I1206 05:40:49.944127 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:49Z","lastTransitionTime":"2025-12-06T05:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.046285 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.046322 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.046334 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.046351 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.046361 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.149067 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.149179 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.149192 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.149210 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.149229 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.252112 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.252156 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.252170 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.252188 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.252203 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.355956 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.356010 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.356021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.356038 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.356050 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.458718 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.458802 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.458826 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.458895 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.458918 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.561919 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.561995 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.562018 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.562044 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.562060 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.662208 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.662358 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.662459 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:50 crc kubenswrapper[4957]: E1206 05:40:50.662594 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:50 crc kubenswrapper[4957]: E1206 05:40:50.662798 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:50 crc kubenswrapper[4957]: E1206 05:40:50.662989 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.665141 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.665244 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.665264 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.665288 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.665308 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.681553 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-slqnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcf7c4c9-894b-450c-83f6-441d3639f75a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgdv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-slqnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.707300 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T05:39:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 05:39:33.204674 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 05:39:33.209541 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3971812975/tls.crt::/tmp/serving-cert-3971812975/tls.key\\\\\\\"\\\\nI1206 05:39:38.756689 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 05:39:38.778597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 05:39:38.778633 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 05:39:38.778669 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 05:39:38.778678 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 05:39:38.804893 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1206 05:39:38.804928 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804935 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 05:39:38.804940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 05:39:38.804944 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 05:39:38.804948 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 05:39:38.804952 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1206 05:39:38.804988 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1206 05:39:38.807114 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.726767 4957 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9019dc7d-622a-4652-86a4-18e5f86931a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T05:39:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9af17c49e0732a21c5ccc770ca8367336ae1b9c225bafebb180a403a5d002d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a3ef434a30109e3d0334417a02c40e10528cbbe8e34e1c2650fdb702bed5ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T05:39:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T05:39:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T05:39:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T05:40:50Z is after 2025-08-24T17:21:41Z" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.789769 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.789914 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.789943 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.789976 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.790001 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.893325 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.893394 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.893405 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.893423 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.893455 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:50Z","lastTransitionTime":"2025-12-06T05:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.926070 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podStartSLOduration=72.92605103 podStartE2EDuration="1m12.92605103s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:50.92603686 +0000 UTC m=+90.576304512" watchObservedRunningTime="2025-12-06 05:40:50.92605103 +0000 UTC m=+90.576318672" Dec 06 05:40:50 crc kubenswrapper[4957]: I1206 05:40:50.969556 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-78cpc" podStartSLOduration=72.969513857 podStartE2EDuration="1m12.969513857s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:50.948151939 +0000 UTC m=+90.598419601" watchObservedRunningTime="2025-12-06 05:40:50.969513857 +0000 UTC m=+90.619781499" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.010112 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.010187 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.010206 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.010232 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.010250 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.021403 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-69r8l" podStartSLOduration=73.021122486 podStartE2EDuration="1m13.021122486s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:50.993701803 +0000 UTC m=+90.643969485" watchObservedRunningTime="2025-12-06 05:40:51.021122486 +0000 UTC m=+90.671390128" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.044210 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pxc62" podStartSLOduration=72.044187081 podStartE2EDuration="1m12.044187081s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:51.02166234 +0000 UTC m=+90.671930012" watchObservedRunningTime="2025-12-06 05:40:51.044187081 +0000 UTC m=+90.694454733" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.044990 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.044980822 podStartE2EDuration="40.044980822s" podCreationTimestamp="2025-12-06 05:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:51.044118829 +0000 UTC m=+90.694386521" watchObservedRunningTime="2025-12-06 05:40:51.044980822 +0000 UTC m=+90.695248464" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.112417 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.112451 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.112462 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.112477 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.112487 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.158053 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.158025745 podStartE2EDuration="1m8.158025745s" podCreationTimestamp="2025-12-06 05:39:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:51.135599757 +0000 UTC m=+90.785867419" watchObservedRunningTime="2025-12-06 05:40:51.158025745 +0000 UTC m=+90.808293387" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.158644 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-g9hn4" podStartSLOduration=73.158634842 podStartE2EDuration="1m13.158634842s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:51.156539975 +0000 UTC m=+90.806807617" watchObservedRunningTime="2025-12-06 05:40:51.158634842 +0000 UTC m=+90.808902494" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.174987 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-jzcdr" podStartSLOduration=72.174965704 podStartE2EDuration="1m12.174965704s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:51.174254035 +0000 UTC m=+90.824521707" watchObservedRunningTime="2025-12-06 05:40:51.174965704 +0000 UTC m=+90.825233356" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.214554 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.214583 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.214594 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.214609 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.214620 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.317479 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.317522 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.317535 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.317552 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.317565 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.421115 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.421184 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.421210 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.421245 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.421271 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.522981 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.523015 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.523026 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.523042 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.523053 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.626086 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.626143 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.626160 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.626184 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.626202 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.662012 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:51 crc kubenswrapper[4957]: E1206 05:40:51.662450 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.728327 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.728388 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.728406 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.728427 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.728443 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.831434 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.831473 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.831482 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.831495 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.831505 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.934566 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.934618 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.934633 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.934651 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:51 crc kubenswrapper[4957]: I1206 05:40:51.934665 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:51Z","lastTransitionTime":"2025-12-06T05:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.037978 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.038025 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.038035 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.038058 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.038074 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.141902 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.141937 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.141945 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.141960 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.141969 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.244328 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.244632 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.244721 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.244812 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.244924 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.347919 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.347997 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.348017 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.348041 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.348058 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.450999 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.451057 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.451074 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.451100 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.451117 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.553571 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.553616 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.553628 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.553645 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.553658 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.656687 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.656731 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.656743 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.656759 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.656769 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.661600 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.661733 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:52 crc kubenswrapper[4957]: E1206 05:40:52.661776 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.661797 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:52 crc kubenswrapper[4957]: E1206 05:40:52.662025 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:52 crc kubenswrapper[4957]: E1206 05:40:52.661942 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.760520 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.760592 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.760612 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.760639 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.760659 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.864134 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.864221 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.864233 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.864259 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.864300 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.967289 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.967351 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.967366 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.967393 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:52 crc kubenswrapper[4957]: I1206 05:40:52.967407 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:52Z","lastTransitionTime":"2025-12-06T05:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.070596 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.070640 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.070652 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.070668 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.070679 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:53Z","lastTransitionTime":"2025-12-06T05:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.173663 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.173737 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.173757 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.173788 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.173811 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:53Z","lastTransitionTime":"2025-12-06T05:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.277672 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.277723 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.277736 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.277755 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.277769 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:53Z","lastTransitionTime":"2025-12-06T05:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.381021 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.381089 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.381115 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.381147 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.381171 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:53Z","lastTransitionTime":"2025-12-06T05:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.484445 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.484783 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.484891 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.484987 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.485068 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:53Z","lastTransitionTime":"2025-12-06T05:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.588701 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.589161 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.589320 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.589458 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.589587 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:53Z","lastTransitionTime":"2025-12-06T05:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.613546 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.613787 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.614022 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.614189 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.614317 4957 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T05:40:53Z","lastTransitionTime":"2025-12-06T05:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.661416 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:53 crc kubenswrapper[4957]: E1206 05:40:53.661608 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.689416 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm"] Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.690985 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.693979 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.694247 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.694379 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.694595 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.739882 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=75.73981197 podStartE2EDuration="1m15.73981197s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:53.739555613 +0000 UTC m=+93.389823265" watchObservedRunningTime="2025-12-06 05:40:53.73981197 +0000 UTC m=+93.390079612" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.768277 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.768339 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.768372 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.768435 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.768463 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.776184 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=19.776154065 podStartE2EDuration="19.776154065s" podCreationTimestamp="2025-12-06 05:40:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:53.758957319 +0000 UTC m=+93.409224961" watchObservedRunningTime="2025-12-06 05:40:53.776154065 +0000 UTC m=+93.426421707" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.869958 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.870020 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.870103 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.870137 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.870166 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.870644 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.871512 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.871700 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.878275 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:53 crc kubenswrapper[4957]: I1206 05:40:53.892751 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0e10bee-2b99-4f6d-b1f3-3d795cd73114-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ll8dm\" (UID: \"c0e10bee-2b99-4f6d-b1f3-3d795cd73114\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:54 crc kubenswrapper[4957]: I1206 05:40:54.007417 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" Dec 06 05:40:54 crc kubenswrapper[4957]: W1206 05:40:54.033747 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e10bee_2b99_4f6d_b1f3_3d795cd73114.slice/crio-1724a8a11c258c1eab2d95a1c5c93294b1ef5d01a8b1c5834233945c0684c1b7 WatchSource:0}: Error finding container 1724a8a11c258c1eab2d95a1c5c93294b1ef5d01a8b1c5834233945c0684c1b7: Status 404 returned error can't find the container with id 1724a8a11c258c1eab2d95a1c5c93294b1ef5d01a8b1c5834233945c0684c1b7 Dec 06 05:40:54 crc kubenswrapper[4957]: I1206 05:40:54.316551 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" event={"ID":"c0e10bee-2b99-4f6d-b1f3-3d795cd73114","Type":"ContainerStarted","Data":"958411fa4c00929e2e05a8bfb84b44a3172dda9102ed7842adabcc20dc1a954f"} Dec 06 05:40:54 crc kubenswrapper[4957]: I1206 05:40:54.318223 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" event={"ID":"c0e10bee-2b99-4f6d-b1f3-3d795cd73114","Type":"ContainerStarted","Data":"1724a8a11c258c1eab2d95a1c5c93294b1ef5d01a8b1c5834233945c0684c1b7"} Dec 06 05:40:54 crc kubenswrapper[4957]: I1206 05:40:54.347090 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ll8dm" podStartSLOduration=76.347056544 podStartE2EDuration="1m16.347056544s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:40:54.344368601 +0000 UTC m=+93.994636283" watchObservedRunningTime="2025-12-06 05:40:54.347056544 +0000 UTC m=+93.997324246" Dec 06 05:40:54 crc kubenswrapper[4957]: I1206 05:40:54.662014 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:54 crc kubenswrapper[4957]: E1206 05:40:54.662234 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:54 crc kubenswrapper[4957]: I1206 05:40:54.662593 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:54 crc kubenswrapper[4957]: E1206 05:40:54.662733 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:54 crc kubenswrapper[4957]: I1206 05:40:54.663175 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:54 crc kubenswrapper[4957]: E1206 05:40:54.663407 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:55 crc kubenswrapper[4957]: I1206 05:40:55.661958 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:55 crc kubenswrapper[4957]: E1206 05:40:55.662196 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:56 crc kubenswrapper[4957]: I1206 05:40:56.662046 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:56 crc kubenswrapper[4957]: E1206 05:40:56.662207 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:56 crc kubenswrapper[4957]: I1206 05:40:56.662310 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:56 crc kubenswrapper[4957]: I1206 05:40:56.662310 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:56 crc kubenswrapper[4957]: E1206 05:40:56.662437 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:56 crc kubenswrapper[4957]: E1206 05:40:56.662527 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:57 crc kubenswrapper[4957]: I1206 05:40:57.662015 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:57 crc kubenswrapper[4957]: E1206 05:40:57.662793 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:40:57 crc kubenswrapper[4957]: I1206 05:40:57.663137 4957 scope.go:117] "RemoveContainer" containerID="4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16" Dec 06 05:40:57 crc kubenswrapper[4957]: E1206 05:40:57.663395 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:40:58 crc kubenswrapper[4957]: I1206 05:40:58.015123 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:58 crc kubenswrapper[4957]: E1206 05:40:58.015355 4957 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:58 crc kubenswrapper[4957]: E1206 05:40:58.015452 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs podName:fcf7c4c9-894b-450c-83f6-441d3639f75a nodeName:}" failed. No retries permitted until 2025-12-06 05:42:02.015429771 +0000 UTC m=+161.665697423 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs") pod "network-metrics-daemon-slqnh" (UID: "fcf7c4c9-894b-450c-83f6-441d3639f75a") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 05:40:58 crc kubenswrapper[4957]: I1206 05:40:58.662064 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:40:58 crc kubenswrapper[4957]: I1206 05:40:58.662185 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:40:58 crc kubenswrapper[4957]: E1206 05:40:58.662237 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:40:58 crc kubenswrapper[4957]: E1206 05:40:58.662394 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:40:58 crc kubenswrapper[4957]: I1206 05:40:58.662532 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:40:58 crc kubenswrapper[4957]: E1206 05:40:58.662716 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:40:59 crc kubenswrapper[4957]: I1206 05:40:59.662081 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:40:59 crc kubenswrapper[4957]: E1206 05:40:59.662303 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:00 crc kubenswrapper[4957]: I1206 05:41:00.661861 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:00 crc kubenswrapper[4957]: I1206 05:41:00.663901 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:00 crc kubenswrapper[4957]: I1206 05:41:00.663959 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:00 crc kubenswrapper[4957]: E1206 05:41:00.664090 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:00 crc kubenswrapper[4957]: E1206 05:41:00.664464 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:00 crc kubenswrapper[4957]: E1206 05:41:00.664638 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:00 crc kubenswrapper[4957]: I1206 05:41:00.685390 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 05:41:01 crc kubenswrapper[4957]: I1206 05:41:01.662297 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:01 crc kubenswrapper[4957]: E1206 05:41:01.663283 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:02 crc kubenswrapper[4957]: I1206 05:41:02.661555 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:02 crc kubenswrapper[4957]: E1206 05:41:02.661786 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:02 crc kubenswrapper[4957]: I1206 05:41:02.661604 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:02 crc kubenswrapper[4957]: I1206 05:41:02.661891 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:02 crc kubenswrapper[4957]: E1206 05:41:02.662040 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:02 crc kubenswrapper[4957]: E1206 05:41:02.662666 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:03 crc kubenswrapper[4957]: I1206 05:41:03.661880 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:03 crc kubenswrapper[4957]: E1206 05:41:03.662095 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:04 crc kubenswrapper[4957]: I1206 05:41:04.661721 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:04 crc kubenswrapper[4957]: I1206 05:41:04.661722 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:04 crc kubenswrapper[4957]: I1206 05:41:04.661908 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:04 crc kubenswrapper[4957]: E1206 05:41:04.662100 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:04 crc kubenswrapper[4957]: E1206 05:41:04.662280 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:04 crc kubenswrapper[4957]: E1206 05:41:04.662425 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:05 crc kubenswrapper[4957]: I1206 05:41:05.661579 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:05 crc kubenswrapper[4957]: E1206 05:41:05.661761 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:06 crc kubenswrapper[4957]: I1206 05:41:06.661994 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:06 crc kubenswrapper[4957]: I1206 05:41:06.662058 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:06 crc kubenswrapper[4957]: I1206 05:41:06.662171 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:06 crc kubenswrapper[4957]: E1206 05:41:06.662242 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:06 crc kubenswrapper[4957]: E1206 05:41:06.662419 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:06 crc kubenswrapper[4957]: E1206 05:41:06.662503 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:07 crc kubenswrapper[4957]: I1206 05:41:07.661441 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:07 crc kubenswrapper[4957]: E1206 05:41:07.661709 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:08 crc kubenswrapper[4957]: I1206 05:41:08.662286 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:08 crc kubenswrapper[4957]: I1206 05:41:08.662326 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:08 crc kubenswrapper[4957]: E1206 05:41:08.662466 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:08 crc kubenswrapper[4957]: I1206 05:41:08.662537 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:08 crc kubenswrapper[4957]: E1206 05:41:08.662698 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:08 crc kubenswrapper[4957]: E1206 05:41:08.662788 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:09 crc kubenswrapper[4957]: I1206 05:41:09.662359 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:09 crc kubenswrapper[4957]: E1206 05:41:09.662912 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:10 crc kubenswrapper[4957]: I1206 05:41:10.661699 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:10 crc kubenswrapper[4957]: I1206 05:41:10.661781 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:10 crc kubenswrapper[4957]: I1206 05:41:10.661899 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:10 crc kubenswrapper[4957]: E1206 05:41:10.664012 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:10 crc kubenswrapper[4957]: E1206 05:41:10.664213 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:10 crc kubenswrapper[4957]: E1206 05:41:10.664369 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:10 crc kubenswrapper[4957]: I1206 05:41:10.704221 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=10.704195702 podStartE2EDuration="10.704195702s" podCreationTimestamp="2025-12-06 05:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:10.702604318 +0000 UTC m=+110.352871980" watchObservedRunningTime="2025-12-06 05:41:10.704195702 +0000 UTC m=+110.354463324" Dec 06 05:41:11 crc kubenswrapper[4957]: I1206 05:41:11.662393 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:11 crc kubenswrapper[4957]: E1206 05:41:11.663125 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.396069 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/1.log" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.396679 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/0.log" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.396724 4957 generic.go:334] "Generic (PLEG): container finished" podID="1384ce42-5e9a-4d6a-a299-f564d1494b0a" containerID="9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41" exitCode=1 Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.396864 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerDied","Data":"9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41"} Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.396953 4957 scope.go:117] "RemoveContainer" containerID="0aec4fe101647dec3db2d25d0958d0fcfad714a9e6e8494fd59f1d971d4b427b" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.397686 4957 scope.go:117] "RemoveContainer" containerID="9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41" Dec 06 05:41:12 crc kubenswrapper[4957]: E1206 05:41:12.398007 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-69r8l_openshift-multus(1384ce42-5e9a-4d6a-a299-f564d1494b0a)\"" pod="openshift-multus/multus-69r8l" podUID="1384ce42-5e9a-4d6a-a299-f564d1494b0a" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.661365 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.661428 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.661389 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:12 crc kubenswrapper[4957]: E1206 05:41:12.661558 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:12 crc kubenswrapper[4957]: E1206 05:41:12.661893 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:12 crc kubenswrapper[4957]: E1206 05:41:12.662383 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:12 crc kubenswrapper[4957]: I1206 05:41:12.662955 4957 scope.go:117] "RemoveContainer" containerID="4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16" Dec 06 05:41:12 crc kubenswrapper[4957]: E1206 05:41:12.663271 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hlm8s_openshift-ovn-kubernetes(fa635ecb-5324-449e-a8f3-7a9bfdca7064)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" Dec 06 05:41:13 crc kubenswrapper[4957]: I1206 05:41:13.404333 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/1.log" Dec 06 05:41:13 crc kubenswrapper[4957]: I1206 05:41:13.661825 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:13 crc kubenswrapper[4957]: E1206 05:41:13.662132 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:14 crc kubenswrapper[4957]: I1206 05:41:14.661506 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:14 crc kubenswrapper[4957]: I1206 05:41:14.661599 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:14 crc kubenswrapper[4957]: E1206 05:41:14.661662 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:14 crc kubenswrapper[4957]: E1206 05:41:14.661766 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:14 crc kubenswrapper[4957]: I1206 05:41:14.662347 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:14 crc kubenswrapper[4957]: E1206 05:41:14.662510 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:15 crc kubenswrapper[4957]: I1206 05:41:15.662174 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:15 crc kubenswrapper[4957]: E1206 05:41:15.662369 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:16 crc kubenswrapper[4957]: I1206 05:41:16.661535 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:16 crc kubenswrapper[4957]: I1206 05:41:16.661696 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:16 crc kubenswrapper[4957]: E1206 05:41:16.661746 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:16 crc kubenswrapper[4957]: I1206 05:41:16.661929 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:16 crc kubenswrapper[4957]: E1206 05:41:16.661826 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:16 crc kubenswrapper[4957]: E1206 05:41:16.662602 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:17 crc kubenswrapper[4957]: I1206 05:41:17.662016 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:17 crc kubenswrapper[4957]: E1206 05:41:17.662126 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:18 crc kubenswrapper[4957]: I1206 05:41:18.661605 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:18 crc kubenswrapper[4957]: I1206 05:41:18.661618 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:18 crc kubenswrapper[4957]: E1206 05:41:18.661810 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:18 crc kubenswrapper[4957]: I1206 05:41:18.661644 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:18 crc kubenswrapper[4957]: E1206 05:41:18.661991 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:18 crc kubenswrapper[4957]: E1206 05:41:18.662138 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:19 crc kubenswrapper[4957]: I1206 05:41:19.661188 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:19 crc kubenswrapper[4957]: E1206 05:41:19.661378 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:20 crc kubenswrapper[4957]: I1206 05:41:20.661407 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:20 crc kubenswrapper[4957]: I1206 05:41:20.661433 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:20 crc kubenswrapper[4957]: E1206 05:41:20.662215 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:20 crc kubenswrapper[4957]: I1206 05:41:20.662262 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:20 crc kubenswrapper[4957]: E1206 05:41:20.662385 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:20 crc kubenswrapper[4957]: E1206 05:41:20.662469 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:20 crc kubenswrapper[4957]: E1206 05:41:20.688608 4957 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 05:41:20 crc kubenswrapper[4957]: E1206 05:41:20.761019 4957 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 05:41:21 crc kubenswrapper[4957]: I1206 05:41:21.662569 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:21 crc kubenswrapper[4957]: E1206 05:41:21.662895 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:22 crc kubenswrapper[4957]: I1206 05:41:22.661598 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:22 crc kubenswrapper[4957]: I1206 05:41:22.661641 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:22 crc kubenswrapper[4957]: E1206 05:41:22.661785 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:22 crc kubenswrapper[4957]: I1206 05:41:22.661951 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:22 crc kubenswrapper[4957]: E1206 05:41:22.663484 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:22 crc kubenswrapper[4957]: E1206 05:41:22.663797 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:23 crc kubenswrapper[4957]: I1206 05:41:23.661362 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:23 crc kubenswrapper[4957]: E1206 05:41:23.661496 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:24 crc kubenswrapper[4957]: I1206 05:41:24.662026 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:24 crc kubenswrapper[4957]: I1206 05:41:24.662153 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:24 crc kubenswrapper[4957]: E1206 05:41:24.662243 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:24 crc kubenswrapper[4957]: I1206 05:41:24.662285 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:24 crc kubenswrapper[4957]: E1206 05:41:24.662455 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:24 crc kubenswrapper[4957]: E1206 05:41:24.662671 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:24 crc kubenswrapper[4957]: I1206 05:41:24.662906 4957 scope.go:117] "RemoveContainer" containerID="9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41" Dec 06 05:41:25 crc kubenswrapper[4957]: I1206 05:41:25.463562 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/1.log" Dec 06 05:41:25 crc kubenswrapper[4957]: I1206 05:41:25.463969 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerStarted","Data":"d2f2175513b8ccbbbf889c9311445008e47014dd16395282f88be03d4396421e"} Dec 06 05:41:25 crc kubenswrapper[4957]: I1206 05:41:25.662222 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:25 crc kubenswrapper[4957]: E1206 05:41:25.662385 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:25 crc kubenswrapper[4957]: I1206 05:41:25.663195 4957 scope.go:117] "RemoveContainer" containerID="4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16" Dec 06 05:41:25 crc kubenswrapper[4957]: E1206 05:41:25.762439 4957 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.471014 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/3.log" Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.474317 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerStarted","Data":"de3f99007b23a6f713f546b61089c454e66fbd1f9502857afebc44f1e5f7f29d"} Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.475549 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.551269 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podStartSLOduration=108.551239979 podStartE2EDuration="1m48.551239979s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:26.513369403 +0000 UTC m=+126.163637035" watchObservedRunningTime="2025-12-06 05:41:26.551239979 +0000 UTC m=+126.201507651" Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.552199 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-slqnh"] Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.552399 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:26 crc kubenswrapper[4957]: E1206 05:41:26.552582 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.662307 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.662398 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:26 crc kubenswrapper[4957]: I1206 05:41:26.662465 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:26 crc kubenswrapper[4957]: E1206 05:41:26.662605 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:26 crc kubenswrapper[4957]: E1206 05:41:26.662781 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:26 crc kubenswrapper[4957]: E1206 05:41:26.662878 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:28 crc kubenswrapper[4957]: I1206 05:41:28.661268 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:28 crc kubenswrapper[4957]: I1206 05:41:28.661345 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:28 crc kubenswrapper[4957]: I1206 05:41:28.661307 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:28 crc kubenswrapper[4957]: E1206 05:41:28.661501 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:28 crc kubenswrapper[4957]: I1206 05:41:28.661518 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:28 crc kubenswrapper[4957]: E1206 05:41:28.661637 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:28 crc kubenswrapper[4957]: E1206 05:41:28.661709 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:28 crc kubenswrapper[4957]: E1206 05:41:28.661761 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:30 crc kubenswrapper[4957]: I1206 05:41:30.662213 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:30 crc kubenswrapper[4957]: I1206 05:41:30.662251 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:30 crc kubenswrapper[4957]: E1206 05:41:30.663250 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 05:41:30 crc kubenswrapper[4957]: I1206 05:41:30.662293 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:30 crc kubenswrapper[4957]: E1206 05:41:30.663371 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 05:41:30 crc kubenswrapper[4957]: I1206 05:41:30.662251 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:30 crc kubenswrapper[4957]: E1206 05:41:30.663496 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 05:41:30 crc kubenswrapper[4957]: E1206 05:41:30.663786 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-slqnh" podUID="fcf7c4c9-894b-450c-83f6-441d3639f75a" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.661689 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.661754 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.661930 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.662274 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.665371 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.665795 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.666092 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.666476 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.669691 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 05:41:32 crc kubenswrapper[4957]: I1206 05:41:32.669750 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.045109 4957 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.081231 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.082328 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.087116 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-284vj"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.087985 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.091695 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.092968 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.093172 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.093374 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.093666 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.093853 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.094076 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.094813 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.094995 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.095225 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6ktjh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.096193 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vlwcn"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.096763 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.096975 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.098098 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.116474 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.116598 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8m4pc"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.116780 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.117065 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.117134 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.117898 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118020 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118212 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118236 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118392 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118477 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118591 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118659 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ftjwd"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.118632 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.119188 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c5ld8"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.119786 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.120174 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ftjwd" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.120539 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vs6qp"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.122098 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.126039 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.126510 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.127108 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.127713 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.128016 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.128405 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.129212 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.130167 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.130985 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.133964 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8mb7l"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.134149 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.134284 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.134467 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.134698 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.134855 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.134988 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.135021 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.135212 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.135387 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.135729 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.135902 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.135958 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.136436 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.136564 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.136721 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.136887 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.138874 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.144108 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.144364 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.144488 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.144639 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.144813 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149007 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149100 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149256 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149381 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149423 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149478 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149553 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149975 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.150157 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.150544 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.151016 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149007 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.151203 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.149384 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.152943 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xlz55"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.153780 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.170977 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173001 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173043 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173207 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173268 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173369 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173445 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173622 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173652 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.173973 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174062 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174145 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174202 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174379 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174464 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174490 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174627 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.174779 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.175366 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.188574 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.189105 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.189371 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.189904 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hj5wv"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.191570 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fls7b"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.191909 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.192051 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jbd4c"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.192439 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.192463 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.193106 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.196691 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vlwcn"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.197222 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.199779 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.201266 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.202044 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.209807 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.210800 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.212765 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.213890 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.214821 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.221305 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.221558 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.221635 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222041 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222225 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gfv9x"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222348 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222769 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222668 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222920 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222639 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.223256 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222716 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.223340 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.222791 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.223816 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.225564 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.226043 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.226084 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.227131 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.230613 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.231188 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.231387 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.231515 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.231605 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.231879 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.236769 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.237148 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.237178 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.237525 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.237188 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.245760 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.246631 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.247229 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.253385 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.253684 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.253909 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.260660 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qrcnq"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.261635 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.265992 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270101 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270166 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28qm5\" (UniqueName: \"kubernetes.io/projected/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-kube-api-access-28qm5\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270260 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270335 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-audit-policies\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270364 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b2983a4e-cc8a-4020-986f-920b7732b840-audit-dir\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270396 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270432 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/da2fdc27-1d40-41fa-b257-57b42be7f811-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-djsdh\" (UID: \"da2fdc27-1d40-41fa-b257-57b42be7f811\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270469 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270516 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270557 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.270881 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-machine-approver-tls\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.271068 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt2xv\" (UniqueName: \"kubernetes.io/projected/c9105dc1-76e5-46dd-854a-c79f1c000c27-kube-api-access-mt2xv\") pod \"downloads-7954f5f757-ftjwd\" (UID: \"c9105dc1-76e5-46dd-854a-c79f1c000c27\") " pod="openshift-console/downloads-7954f5f757-ftjwd" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.271409 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-audit-policies\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272188 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272393 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-serving-cert\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272487 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272567 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272623 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272701 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf99g\" (UniqueName: \"kubernetes.io/projected/70c2a386-5150-4abb-9e97-af87740cc325-kube-api-access-sf99g\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272754 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272795 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272913 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.272972 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-auth-proxy-config\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.273027 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70c2a386-5150-4abb-9e97-af87740cc325-audit-dir\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.273061 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.273108 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27lcg\" (UniqueName: \"kubernetes.io/projected/b2983a4e-cc8a-4020-986f-920b7732b840-kube-api-access-27lcg\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.273152 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.273197 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-config\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.274066 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxvn\" (UniqueName: \"kubernetes.io/projected/da2fdc27-1d40-41fa-b257-57b42be7f811-kube-api-access-bnxvn\") pod \"cluster-samples-operator-665b6dd947-djsdh\" (UID: \"da2fdc27-1d40-41fa-b257-57b42be7f811\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.274309 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-encryption-config\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.274388 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.274483 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-etcd-client\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.274577 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmfwg\" (UniqueName: \"kubernetes.io/projected/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-kube-api-access-bmfwg\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.276782 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.276859 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.277209 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.281609 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.285243 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.292584 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qkg8t"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.294073 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.294239 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.294589 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.294860 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.295756 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.300896 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.301766 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.304208 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.304975 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.306572 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.308550 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9kjdh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.309577 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.314053 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.315035 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.319927 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.320575 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.322218 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.324695 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8m4pc"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.325976 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ftjwd"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.326507 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.327967 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.330002 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fls7b"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.331952 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-tbzz4"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.334593 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.334705 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.336693 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.342001 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-284vj"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.342093 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xlz55"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.343486 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.345681 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.345947 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.347154 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jbd4c"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.348324 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.349345 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gfv9x"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.350727 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.351862 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.354508 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c5ld8"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.354636 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.358244 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.359529 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vs6qp"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.363700 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.364728 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.364780 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.367172 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.367215 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.370046 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hj5wv"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.370135 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nlpbn"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.371024 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.371149 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.372374 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6ktjh"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.373337 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.375755 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nlpbn"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.376156 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qkg8t"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378560 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378636 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf99g\" (UniqueName: \"kubernetes.io/projected/70c2a386-5150-4abb-9e97-af87740cc325-kube-api-access-sf99g\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378677 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378709 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378779 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378806 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-auth-proxy-config\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378855 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70c2a386-5150-4abb-9e97-af87740cc325-audit-dir\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378891 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378918 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27lcg\" (UniqueName: \"kubernetes.io/projected/b2983a4e-cc8a-4020-986f-920b7732b840-kube-api-access-27lcg\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378942 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.378975 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-config\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379011 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxvn\" (UniqueName: \"kubernetes.io/projected/da2fdc27-1d40-41fa-b257-57b42be7f811-kube-api-access-bnxvn\") pod \"cluster-samples-operator-665b6dd947-djsdh\" (UID: \"da2fdc27-1d40-41fa-b257-57b42be7f811\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379047 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-encryption-config\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379073 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379110 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-etcd-client\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379152 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmfwg\" (UniqueName: \"kubernetes.io/projected/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-kube-api-access-bmfwg\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379188 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379221 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379248 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28qm5\" (UniqueName: \"kubernetes.io/projected/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-kube-api-access-28qm5\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379279 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379284 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tbzz4"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379311 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-audit-policies\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379336 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b2983a4e-cc8a-4020-986f-920b7732b840-audit-dir\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379366 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379440 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/da2fdc27-1d40-41fa-b257-57b42be7f811-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-djsdh\" (UID: \"da2fdc27-1d40-41fa-b257-57b42be7f811\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379468 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379501 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379527 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379597 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-machine-approver-tls\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379624 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt2xv\" (UniqueName: \"kubernetes.io/projected/c9105dc1-76e5-46dd-854a-c79f1c000c27-kube-api-access-mt2xv\") pod \"downloads-7954f5f757-ftjwd\" (UID: \"c9105dc1-76e5-46dd-854a-c79f1c000c27\") " pod="openshift-console/downloads-7954f5f757-ftjwd" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379656 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-audit-policies\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379700 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-serving-cert\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379725 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379763 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.379919 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70c2a386-5150-4abb-9e97-af87740cc325-audit-dir\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.380483 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-auth-proxy-config\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.380943 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.381716 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qrcnq"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.384024 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-audit-policies\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.384218 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.385030 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-config\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.386608 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.386768 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b2983a4e-cc8a-4020-986f-920b7732b840-audit-dir\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.388780 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.389226 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.390305 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.390320 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.390715 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.391145 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.391089 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-audit-policies\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.391289 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.391340 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.392538 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/da2fdc27-1d40-41fa-b257-57b42be7f811-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-djsdh\" (UID: \"da2fdc27-1d40-41fa-b257-57b42be7f811\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.393193 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.393481 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.394159 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2983a4e-cc8a-4020-986f-920b7732b840-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.394300 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-serving-cert\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.394727 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.395165 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8mb7l"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.395978 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-machine-approver-tls\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.396178 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.396444 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-encryption-config\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.397686 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.397733 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.398547 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b2983a4e-cc8a-4020-986f-920b7732b840-etcd-client\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.399909 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.406260 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.402352 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.406691 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5g4vs"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.407499 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.407521 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9zmkv"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.409020 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.411241 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9zmkv"] Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.425300 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.444635 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.464613 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.485749 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.505330 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.524996 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.545255 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.564872 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.584793 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.605262 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.645824 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.646920 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.665601 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.685432 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.705120 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.725081 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.744634 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.764421 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.785439 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.804476 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.824637 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.844478 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.864285 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.904679 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.925101 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.952638 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.964767 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.985122 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.986476 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-registry-certificates\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.986632 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qkw4\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-kube-api-access-6qkw4\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.986787 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-config\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.986933 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-serving-cert\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987227 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-available-featuregates\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987291 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kk7b\" (UniqueName: \"kubernetes.io/projected/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-kube-api-access-7kk7b\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987352 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-bound-sa-token\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987392 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-registry-tls\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987424 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-trusted-ca\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987456 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-serving-cert\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987488 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-config\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987520 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-service-ca\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987567 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987617 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e31a64-6ce7-4b24-9891-9775e7193b8a-serving-cert\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987694 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9ad5cbf-adfe-4012-8968-0d7748db277d-images\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987740 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fprz\" (UniqueName: \"kubernetes.io/projected/b9e31a64-6ce7-4b24-9891-9775e7193b8a-kube-api-access-5fprz\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987887 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vkxr\" (UniqueName: \"kubernetes.io/projected/950d1218-8492-4736-8dbb-cfa12d544bae-kube-api-access-6vkxr\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.987990 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-trusted-ca-bundle\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.988056 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-config\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.988136 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c269452-c147-4adc-9280-7dd91cff2687-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: E1206 05:41:34.988255 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.488149841 +0000 UTC m=+135.138417483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.988363 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghrg4\" (UniqueName: \"kubernetes.io/projected/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-kube-api-access-ghrg4\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.988466 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdpww\" (UniqueName: \"kubernetes.io/projected/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-kube-api-access-bdpww\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.988574 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.988677 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-serving-cert\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.988854 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-config\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.989050 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-client-ca\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.989164 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9ad5cbf-adfe-4012-8968-0d7748db277d-config\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.989336 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950d1218-8492-4736-8dbb-cfa12d544bae-serving-cert\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.989445 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-trusted-ca\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.989565 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-serving-cert\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.989816 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-service-ca-bundle\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.989922 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-oauth-serving-cert\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990013 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c269452-c147-4adc-9280-7dd91cff2687-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990046 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swrrj\" (UniqueName: \"kubernetes.io/projected/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-kube-api-access-swrrj\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990118 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-client-ca\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990263 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ad5cbf-adfe-4012-8968-0d7748db277d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990863 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-oauth-config\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990916 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsxnw\" (UniqueName: \"kubernetes.io/projected/b9ad5cbf-adfe-4012-8968-0d7748db277d-kube-api-access-qsxnw\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990950 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:34 crc kubenswrapper[4957]: I1206 05:41:34.990988 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-config\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.025393 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.044589 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.066880 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.085075 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093460 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093627 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqf7m\" (UniqueName: \"kubernetes.io/projected/fa307629-37e4-43eb-bd29-74dda3004967-kube-api-access-fqf7m\") pod \"migrator-59844c95c7-9ckdj\" (UID: \"fa307629-37e4-43eb-bd29-74dda3004967\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.093650 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.593628955 +0000 UTC m=+135.243896597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093676 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7665fa45-3b81-48a1-ba82-e6471ae634e9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093710 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hdb8\" (UniqueName: \"kubernetes.io/projected/26e1869e-3f67-4f4d-8aad-3007fbdeb4ec-kube-api-access-4hdb8\") pod \"multus-admission-controller-857f4d67dd-qkg8t\" (UID: \"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093765 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c269452-c147-4adc-9280-7dd91cff2687-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093787 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghrg4\" (UniqueName: \"kubernetes.io/projected/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-kube-api-access-ghrg4\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093806 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093899 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9af915f-4e5b-440d-a3b3-534d16b1f686-secret-volume\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093933 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/72514156-b413-4c84-802b-f114ebed8090-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9xm5h\" (UID: \"72514156-b413-4c84-802b-f114ebed8090\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.093995 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-service-ca\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.094029 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eddcefaa-1fba-4c03-8b33-886c131532e0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.094050 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/34af5be8-778f-43a8-8241-3ece02457a25-profile-collector-cert\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.094170 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd4nk\" (UniqueName: \"kubernetes.io/projected/7c4ad024-e034-4584-a801-7f59d97e588f-kube-api-access-gd4nk\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.094377 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/922c58c7-e5f1-4e55-bf25-a903738bea1b-metrics-tls\") pod \"dns-operator-744455d44c-fls7b\" (UID: \"922c58c7-e5f1-4e55-bf25-a903738bea1b\") " pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095063 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eddcefaa-1fba-4c03-8b33-886c131532e0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095111 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9ef9416-d92a-401b-8f80-81ab25b24d31-service-ca-bundle\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095169 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4ad024-e034-4584-a801-7f59d97e588f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095209 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-encryption-config\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095248 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66cmp\" (UniqueName: \"kubernetes.io/projected/92ffc396-7acd-4dd1-8bff-61901a7e6159-kube-api-access-66cmp\") pod \"control-plane-machine-set-operator-78cbb6b69f-m7xn8\" (UID: \"92ffc396-7acd-4dd1-8bff-61901a7e6159\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095289 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-client-ca\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095300 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095332 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8498422c-9c5b-46cc-a19b-4c381511abf0-metrics-tls\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095371 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-serving-cert\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095409 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950d1218-8492-4736-8dbb-cfa12d544bae-serving-cert\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095879 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpknx\" (UniqueName: \"kubernetes.io/projected/665360fe-3cbc-49bb-9316-01a6cf62174e-kube-api-access-fpknx\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095926 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686zt\" (UniqueName: \"kubernetes.io/projected/aa443eb3-221f-42d3-a946-f5bfee0d7148-kube-api-access-686zt\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.095957 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-oauth-serving-cert\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.096125 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtrf8\" (UniqueName: \"kubernetes.io/projected/34af5be8-778f-43a8-8241-3ece02457a25-kube-api-access-jtrf8\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.096197 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-socket-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.096241 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c269452-c147-4adc-9280-7dd91cff2687-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.096267 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w7tl\" (UniqueName: \"kubernetes.io/projected/8498422c-9c5b-46cc-a19b-4c381511abf0-kube-api-access-2w7tl\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.096311 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-certs\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.096798 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aaeb2a27-18ff-4750-bd39-e24c8e59665a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.096824 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c269452-c147-4adc-9280-7dd91cff2687-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097021 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ad5cbf-adfe-4012-8968-0d7748db277d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097136 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097162 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-oauth-serving-cert\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097186 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzd8n\" (UniqueName: \"kubernetes.io/projected/37a32874-c303-4090-b86e-3d2caf34d36a-kube-api-access-tzd8n\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097222 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zndq5\" (UniqueName: \"kubernetes.io/projected/297926a5-4372-4868-a1b8-3052078d4217-kube-api-access-zndq5\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097272 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097338 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/297926a5-4372-4868-a1b8-3052078d4217-proxy-tls\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097381 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-config\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097417 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097461 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37a32874-c303-4090-b86e-3d2caf34d36a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097513 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a84a237c-277b-43f3-82ab-6c277059c76c-signing-key\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097556 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-mountpoint-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097611 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-srv-cert\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097658 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qkw4\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-kube-api-access-6qkw4\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097700 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-serving-cert\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.097772 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-available-featuregates\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.099578 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-plugins-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.099764 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-bound-sa-token\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.099866 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhz4k\" (UniqueName: \"kubernetes.io/projected/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-kube-api-access-bhz4k\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.099963 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/37a32874-c303-4090-b86e-3d2caf34d36a-images\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100052 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aaeb2a27-18ff-4750-bd39-e24c8e59665a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100182 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-serving-cert\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100265 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbhmm\" (UniqueName: \"kubernetes.io/projected/f9ef9416-d92a-401b-8f80-81ab25b24d31-kube-api-access-gbhmm\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100336 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62sl\" (UniqueName: \"kubernetes.io/projected/257d6f9b-ac33-4f18-ab46-2ada42beb644-kube-api-access-d62sl\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100421 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-etcd-client\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100522 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100613 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-metrics-certs\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100700 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7665fa45-3b81-48a1-ba82-e6471ae634e9-config\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100797 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-node-bootstrap-token\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100899 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-registration-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100974 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd7ch\" (UniqueName: \"kubernetes.io/projected/dec5d39a-fa0a-413c-b4a9-014d95bae8e0-kube-api-access-fd7ch\") pod \"ingress-canary-nlpbn\" (UID: \"dec5d39a-fa0a-413c-b4a9-014d95bae8e0\") " pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100932 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950d1218-8492-4736-8dbb-cfa12d544bae-serving-cert\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100177 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.099880 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-client-ca\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.100973 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-serving-cert\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101204 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82xbl\" (UniqueName: \"kubernetes.io/projected/1f2412f9-fb1d-49fa-9337-d9b526ce074d-kube-api-access-82xbl\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101366 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9af915f-4e5b-440d-a3b3-534d16b1f686-config-volume\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101431 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-trusted-ca-bundle\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.101453 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.601432065 +0000 UTC m=+135.251699707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101504 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-config\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101557 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf7tj\" (UniqueName: \"kubernetes.io/projected/a84a237c-277b-43f3-82ab-6c277059c76c-kube-api-access-tf7tj\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101605 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12d0b3b5-412a-42a9-8666-60d47698a3e7-metrics-tls\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101647 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e299c820-62cc-4e26-90fc-d80670866a0a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101686 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a95ad0d-9149-4c33-9f57-a9ca76143509-serving-cert\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.101757 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-image-import-ca\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.102388 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/92ffc396-7acd-4dd1-8bff-61901a7e6159-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m7xn8\" (UID: \"92ffc396-7acd-4dd1-8bff-61901a7e6159\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.102473 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqnl5\" (UniqueName: \"kubernetes.io/projected/922c58c7-e5f1-4e55-bf25-a903738bea1b-kube-api-access-vqnl5\") pod \"dns-operator-744455d44c-fls7b\" (UID: \"922c58c7-e5f1-4e55-bf25-a903738bea1b\") " pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.102600 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdpww\" (UniqueName: \"kubernetes.io/projected/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-kube-api-access-bdpww\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.102678 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-serving-cert\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.102763 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5lh4\" (UniqueName: \"kubernetes.io/projected/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-kube-api-access-k5lh4\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.102885 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-serving-cert\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.102952 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103041 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-config\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103076 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-client\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103137 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-config\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103200 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/297926a5-4372-4868-a1b8-3052078d4217-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103237 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c4ad024-e034-4584-a801-7f59d97e588f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103295 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-default-certificate\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103327 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/257d6f9b-ac33-4f18-ab46-2ada42beb644-apiservice-cert\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103401 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v96m\" (UniqueName: \"kubernetes.io/projected/eddcefaa-1fba-4c03-8b33-886c131532e0-kube-api-access-6v96m\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103439 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9ad5cbf-adfe-4012-8968-0d7748db277d-config\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103473 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ps2f\" (UniqueName: \"kubernetes.io/projected/f9af915f-4e5b-440d-a3b3-534d16b1f686-kube-api-access-4ps2f\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103508 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq8lb\" (UniqueName: \"kubernetes.io/projected/12d0b3b5-412a-42a9-8666-60d47698a3e7-kube-api-access-zq8lb\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103542 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-trusted-ca\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103574 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-service-ca-bundle\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103607 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f2412f9-fb1d-49fa-9337-d9b526ce074d-serving-cert\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103638 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.103664 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-config\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.104339 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-trusted-ca-bundle\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.105009 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-config\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.105717 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.106247 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swrrj\" (UniqueName: \"kubernetes.io/projected/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-kube-api-access-swrrj\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.106422 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-config\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.106525 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8498422c-9c5b-46cc-a19b-4c381511abf0-config-volume\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.107717 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-service-ca-bundle\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.107886 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a95ad0d-9149-4c33-9f57-a9ca76143509-config\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.107804 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-config\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.108367 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-client-ca\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.108472 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.108375 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-serving-cert\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.108378 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9ad5cbf-adfe-4012-8968-0d7748db277d-config\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.108820 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-trusted-ca\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.108962 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7665fa45-3b81-48a1-ba82-e6471ae634e9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.108975 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-available-featuregates\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.109074 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dec5d39a-fa0a-413c-b4a9-014d95bae8e0-cert\") pod \"ingress-canary-nlpbn\" (UID: \"dec5d39a-fa0a-413c-b4a9-014d95bae8e0\") " pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.109470 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/26e1869e-3f67-4f4d-8aad-3007fbdeb4ec-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qkg8t\" (UID: \"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.109545 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-client-ca\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.109656 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.109742 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37a32874-c303-4090-b86e-3d2caf34d36a-proxy-tls\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.109845 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jznsp\" (UniqueName: \"kubernetes.io/projected/9a95ad0d-9149-4c33-9f57-a9ca76143509-kube-api-access-jznsp\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110040 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-oauth-config\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110132 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e299c820-62cc-4e26-90fc-d80670866a0a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110213 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsxnw\" (UniqueName: \"kubernetes.io/projected/b9ad5cbf-adfe-4012-8968-0d7748db277d-kube-api-access-qsxnw\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110312 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-audit\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110398 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt8lt\" (UniqueName: \"kubernetes.io/projected/cd29b564-939e-4304-9f8e-11aefcb02a5e-kube-api-access-qt8lt\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110477 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/34af5be8-778f-43a8-8241-3ece02457a25-srv-cert\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110568 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-ca\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110644 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cd29b564-939e-4304-9f8e-11aefcb02a5e-node-pullsecrets\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110916 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd29b564-939e-4304-9f8e-11aefcb02a5e-audit-dir\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110995 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaeb2a27-18ff-4750-bd39-e24c8e59665a-config\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111086 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-registry-certificates\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111159 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-config\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111233 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kk7b\" (UniqueName: \"kubernetes.io/projected/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-kube-api-access-7kk7b\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111320 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a84a237c-277b-43f3-82ab-6c277059c76c-signing-cabundle\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111399 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-csi-data-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111476 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12d0b3b5-412a-42a9-8666-60d47698a3e7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111559 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e299c820-62cc-4e26-90fc-d80670866a0a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111633 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7nrq\" (UniqueName: \"kubernetes.io/projected/574e7efb-63cb-4cbe-a41c-abc7e55d165f-kube-api-access-j7nrq\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111736 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-registry-tls\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111816 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-trusted-ca\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111915 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/257d6f9b-ac33-4f18-ab46-2ada42beb644-tmpfs\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.111999 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sg89\" (UniqueName: \"kubernetes.io/projected/72514156-b413-4c84-802b-f114ebed8090-kube-api-access-4sg89\") pod \"package-server-manager-789f6589d5-9xm5h\" (UID: \"72514156-b413-4c84-802b-f114ebed8090\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112125 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-config\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112226 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-service-ca\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112392 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-stats-auth\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112523 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e31a64-6ce7-4b24-9891-9775e7193b8a-serving-cert\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112638 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12d0b3b5-412a-42a9-8666-60d47698a3e7-trusted-ca\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112878 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9ad5cbf-adfe-4012-8968-0d7748db277d-images\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112987 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-etcd-serving-ca\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.113088 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vkxr\" (UniqueName: \"kubernetes.io/projected/950d1218-8492-4736-8dbb-cfa12d544bae-kube-api-access-6vkxr\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.113168 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-config\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.113289 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fprz\" (UniqueName: \"kubernetes.io/projected/b9e31a64-6ce7-4b24-9891-9775e7193b8a-kube-api-access-5fprz\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.113372 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/257d6f9b-ac33-4f18-ab46-2ada42beb644-webhook-cert\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.110485 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-serving-cert\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.112425 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-registry-certificates\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.114381 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-serving-cert\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.114780 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9ad5cbf-adfe-4012-8968-0d7748db277d-images\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.114805 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-config\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.115094 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-trusted-ca\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.115146 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c269452-c147-4adc-9280-7dd91cff2687-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.115346 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-service-ca\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.117008 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ad5cbf-adfe-4012-8968-0d7748db277d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.118280 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e31a64-6ce7-4b24-9891-9775e7193b8a-serving-cert\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.119315 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-oauth-config\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.120855 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-registry-tls\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.126322 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.144322 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.165486 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.185509 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.206070 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.214786 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.215040 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.714996625 +0000 UTC m=+135.365264257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.215127 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v96m\" (UniqueName: \"kubernetes.io/projected/eddcefaa-1fba-4c03-8b33-886c131532e0-kube-api-access-6v96m\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.215197 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ps2f\" (UniqueName: \"kubernetes.io/projected/f9af915f-4e5b-440d-a3b3-534d16b1f686-kube-api-access-4ps2f\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.215231 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq8lb\" (UniqueName: \"kubernetes.io/projected/12d0b3b5-412a-42a9-8666-60d47698a3e7-kube-api-access-zq8lb\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.215282 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f2412f9-fb1d-49fa-9337-d9b526ce074d-serving-cert\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.215378 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216177 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-config\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216237 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8498422c-9c5b-46cc-a19b-4c381511abf0-config-volume\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216267 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a95ad0d-9149-4c33-9f57-a9ca76143509-config\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216377 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216460 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7665fa45-3b81-48a1-ba82-e6471ae634e9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216494 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dec5d39a-fa0a-413c-b4a9-014d95bae8e0-cert\") pod \"ingress-canary-nlpbn\" (UID: \"dec5d39a-fa0a-413c-b4a9-014d95bae8e0\") " pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216536 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/26e1869e-3f67-4f4d-8aad-3007fbdeb4ec-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qkg8t\" (UID: \"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216571 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216601 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37a32874-c303-4090-b86e-3d2caf34d36a-proxy-tls\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216624 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jznsp\" (UniqueName: \"kubernetes.io/projected/9a95ad0d-9149-4c33-9f57-a9ca76143509-kube-api-access-jznsp\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216660 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e299c820-62cc-4e26-90fc-d80670866a0a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216696 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-audit\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216716 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt8lt\" (UniqueName: \"kubernetes.io/projected/cd29b564-939e-4304-9f8e-11aefcb02a5e-kube-api-access-qt8lt\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216744 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/34af5be8-778f-43a8-8241-3ece02457a25-srv-cert\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216771 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-ca\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216793 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cd29b564-939e-4304-9f8e-11aefcb02a5e-node-pullsecrets\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216816 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd29b564-939e-4304-9f8e-11aefcb02a5e-audit-dir\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216874 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaeb2a27-18ff-4750-bd39-e24c8e59665a-config\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216959 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a84a237c-277b-43f3-82ab-6c277059c76c-signing-cabundle\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.216990 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-csi-data-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217032 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12d0b3b5-412a-42a9-8666-60d47698a3e7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217073 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e299c820-62cc-4e26-90fc-d80670866a0a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217118 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7nrq\" (UniqueName: \"kubernetes.io/projected/574e7efb-63cb-4cbe-a41c-abc7e55d165f-kube-api-access-j7nrq\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217167 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/257d6f9b-ac33-4f18-ab46-2ada42beb644-tmpfs\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217190 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sg89\" (UniqueName: \"kubernetes.io/projected/72514156-b413-4c84-802b-f114ebed8090-kube-api-access-4sg89\") pod \"package-server-manager-789f6589d5-9xm5h\" (UID: \"72514156-b413-4c84-802b-f114ebed8090\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217213 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-stats-auth\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217236 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12d0b3b5-412a-42a9-8666-60d47698a3e7-trusted-ca\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217272 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-etcd-serving-ca\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217308 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/257d6f9b-ac33-4f18-ab46-2ada42beb644-webhook-cert\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217373 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqf7m\" (UniqueName: \"kubernetes.io/projected/fa307629-37e4-43eb-bd29-74dda3004967-kube-api-access-fqf7m\") pod \"migrator-59844c95c7-9ckdj\" (UID: \"fa307629-37e4-43eb-bd29-74dda3004967\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217410 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7665fa45-3b81-48a1-ba82-e6471ae634e9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217434 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hdb8\" (UniqueName: \"kubernetes.io/projected/26e1869e-3f67-4f4d-8aad-3007fbdeb4ec-kube-api-access-4hdb8\") pod \"multus-admission-controller-857f4d67dd-qkg8t\" (UID: \"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217480 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9af915f-4e5b-440d-a3b3-534d16b1f686-secret-volume\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217505 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/72514156-b413-4c84-802b-f114ebed8090-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9xm5h\" (UID: \"72514156-b413-4c84-802b-f114ebed8090\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217530 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-service-ca\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217548 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eddcefaa-1fba-4c03-8b33-886c131532e0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217567 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/34af5be8-778f-43a8-8241-3ece02457a25-profile-collector-cert\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217589 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd4nk\" (UniqueName: \"kubernetes.io/projected/7c4ad024-e034-4584-a801-7f59d97e588f-kube-api-access-gd4nk\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217610 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/922c58c7-e5f1-4e55-bf25-a903738bea1b-metrics-tls\") pod \"dns-operator-744455d44c-fls7b\" (UID: \"922c58c7-e5f1-4e55-bf25-a903738bea1b\") " pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217638 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eddcefaa-1fba-4c03-8b33-886c131532e0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217658 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9ef9416-d92a-401b-8f80-81ab25b24d31-service-ca-bundle\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217688 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4ad024-e034-4584-a801-7f59d97e588f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217717 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-encryption-config\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217742 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66cmp\" (UniqueName: \"kubernetes.io/projected/92ffc396-7acd-4dd1-8bff-61901a7e6159-kube-api-access-66cmp\") pod \"control-plane-machine-set-operator-78cbb6b69f-m7xn8\" (UID: \"92ffc396-7acd-4dd1-8bff-61901a7e6159\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217773 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8498422c-9c5b-46cc-a19b-4c381511abf0-metrics-tls\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217801 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpknx\" (UniqueName: \"kubernetes.io/projected/665360fe-3cbc-49bb-9316-01a6cf62174e-kube-api-access-fpknx\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217822 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686zt\" (UniqueName: \"kubernetes.io/projected/aa443eb3-221f-42d3-a946-f5bfee0d7148-kube-api-access-686zt\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217884 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtrf8\" (UniqueName: \"kubernetes.io/projected/34af5be8-778f-43a8-8241-3ece02457a25-kube-api-access-jtrf8\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217894 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a95ad0d-9149-4c33-9f57-a9ca76143509-config\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217910 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-socket-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218023 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w7tl\" (UniqueName: \"kubernetes.io/projected/8498422c-9c5b-46cc-a19b-4c381511abf0-kube-api-access-2w7tl\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218109 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-certs\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218151 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aaeb2a27-18ff-4750-bd39-e24c8e59665a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218232 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218283 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzd8n\" (UniqueName: \"kubernetes.io/projected/37a32874-c303-4090-b86e-3d2caf34d36a-kube-api-access-tzd8n\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218305 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218325 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-socket-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218336 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zndq5\" (UniqueName: \"kubernetes.io/projected/297926a5-4372-4868-a1b8-3052078d4217-kube-api-access-zndq5\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218424 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e299c820-62cc-4e26-90fc-d80670866a0a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.219193 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-audit\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.219593 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-service-ca\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220241 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.217422 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-config\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220400 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-etcd-serving-ca\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220492 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f2412f9-fb1d-49fa-9337-d9b526ce074d-serving-cert\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220596 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-ca\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220714 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/cd29b564-939e-4304-9f8e-11aefcb02a5e-node-pullsecrets\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220783 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cd29b564-939e-4304-9f8e-11aefcb02a5e-audit-dir\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.218469 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/297926a5-4372-4868-a1b8-3052078d4217-proxy-tls\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220945 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.220990 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37a32874-c303-4090-b86e-3d2caf34d36a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221012 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-csi-data-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221046 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a84a237c-277b-43f3-82ab-6c277059c76c-signing-key\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221076 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-mountpoint-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221099 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-srv-cert\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221138 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-plugins-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221166 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhz4k\" (UniqueName: \"kubernetes.io/projected/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-kube-api-access-bhz4k\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221186 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/37a32874-c303-4090-b86e-3d2caf34d36a-images\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221210 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aaeb2a27-18ff-4750-bd39-e24c8e59665a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221232 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbhmm\" (UniqueName: \"kubernetes.io/projected/f9ef9416-d92a-401b-8f80-81ab25b24d31-kube-api-access-gbhmm\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221233 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-mountpoint-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221254 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62sl\" (UniqueName: \"kubernetes.io/projected/257d6f9b-ac33-4f18-ab46-2ada42beb644-kube-api-access-d62sl\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221278 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-etcd-client\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221299 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221320 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-metrics-certs\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221339 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7665fa45-3b81-48a1-ba82-e6471ae634e9-config\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221369 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-node-bootstrap-token\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221388 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-registration-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221402 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/257d6f9b-ac33-4f18-ab46-2ada42beb644-tmpfs\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221482 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd7ch\" (UniqueName: \"kubernetes.io/projected/dec5d39a-fa0a-413c-b4a9-014d95bae8e0-kube-api-access-fd7ch\") pod \"ingress-canary-nlpbn\" (UID: \"dec5d39a-fa0a-413c-b4a9-014d95bae8e0\") " pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221505 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82xbl\" (UniqueName: \"kubernetes.io/projected/1f2412f9-fb1d-49fa-9337-d9b526ce074d-kube-api-access-82xbl\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221523 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9af915f-4e5b-440d-a3b3-534d16b1f686-config-volume\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221546 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf7tj\" (UniqueName: \"kubernetes.io/projected/a84a237c-277b-43f3-82ab-6c277059c76c-kube-api-access-tf7tj\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221564 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12d0b3b5-412a-42a9-8666-60d47698a3e7-metrics-tls\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221581 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e299c820-62cc-4e26-90fc-d80670866a0a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221599 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a95ad0d-9149-4c33-9f57-a9ca76143509-serving-cert\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221658 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-image-import-ca\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221686 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/92ffc396-7acd-4dd1-8bff-61901a7e6159-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m7xn8\" (UID: \"92ffc396-7acd-4dd1-8bff-61901a7e6159\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221711 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqnl5\" (UniqueName: \"kubernetes.io/projected/922c58c7-e5f1-4e55-bf25-a903738bea1b-kube-api-access-vqnl5\") pod \"dns-operator-744455d44c-fls7b\" (UID: \"922c58c7-e5f1-4e55-bf25-a903738bea1b\") " pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221746 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-serving-cert\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221777 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5lh4\" (UniqueName: \"kubernetes.io/projected/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-kube-api-access-k5lh4\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221804 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221848 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-client\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221868 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-config\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221888 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/297926a5-4372-4868-a1b8-3052078d4217-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221911 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c4ad024-e034-4584-a801-7f59d97e588f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221935 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-default-certificate\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.221980 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/257d6f9b-ac33-4f18-ab46-2ada42beb644-apiservice-cert\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.222154 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37a32874-c303-4090-b86e-3d2caf34d36a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.222288 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12d0b3b5-412a-42a9-8666-60d47698a3e7-trusted-ca\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.222368 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-plugins-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.222529 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.222819 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa443eb3-221f-42d3-a946-f5bfee0d7148-registration-dir\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.223607 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c4ad024-e034-4584-a801-7f59d97e588f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.223732 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4ad024-e034-4584-a801-7f59d97e588f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.224110 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.724093163 +0000 UTC m=+135.374360795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.224538 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/922c58c7-e5f1-4e55-bf25-a903738bea1b-metrics-tls\") pod \"dns-operator-744455d44c-fls7b\" (UID: \"922c58c7-e5f1-4e55-bf25-a903738bea1b\") " pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.224604 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7665fa45-3b81-48a1-ba82-e6471ae634e9-config\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.224550 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-config\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.224561 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/297926a5-4372-4868-a1b8-3052078d4217-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.226000 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.226761 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7665fa45-3b81-48a1-ba82-e6471ae634e9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.227471 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/72514156-b413-4c84-802b-f114ebed8090-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9xm5h\" (UID: \"72514156-b413-4c84-802b-f114ebed8090\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.227601 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/cd29b564-939e-4304-9f8e-11aefcb02a5e-image-import-ca\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.227649 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e299c820-62cc-4e26-90fc-d80670866a0a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.228186 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.229178 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-serving-cert\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.230244 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.231219 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12d0b3b5-412a-42a9-8666-60d47698a3e7-metrics-tls\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.231878 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a95ad0d-9149-4c33-9f57-a9ca76143509-serving-cert\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.232267 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-etcd-client\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.232404 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1f2412f9-fb1d-49fa-9337-d9b526ce074d-etcd-client\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.232636 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/92ffc396-7acd-4dd1-8bff-61901a7e6159-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m7xn8\" (UID: \"92ffc396-7acd-4dd1-8bff-61901a7e6159\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.233311 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/cd29b564-939e-4304-9f8e-11aefcb02a5e-encryption-config\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.245735 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.263791 4957 request.go:700] Waited for 1.000920823s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/secrets?fieldSelector=metadata.name%3Dsigning-key&limit=500&resourceVersion=0 Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.266187 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.276725 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a84a237c-277b-43f3-82ab-6c277059c76c-signing-key\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.284261 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.292421 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a84a237c-277b-43f3-82ab-6c277059c76c-signing-cabundle\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.306908 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.322759 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.323034 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.823002 +0000 UTC m=+135.473269662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.323679 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.323694 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.324036 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.824024299 +0000 UTC m=+135.474291921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.333499 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/37a32874-c303-4090-b86e-3d2caf34d36a-images\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.345056 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.365540 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.372559 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37a32874-c303-4090-b86e-3d2caf34d36a-proxy-tls\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.385862 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.392142 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aaeb2a27-18ff-4750-bd39-e24c8e59665a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.405391 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.425578 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.426091 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.427350 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:35.92732008 +0000 UTC m=+135.577587712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.432767 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaeb2a27-18ff-4750-bd39-e24c8e59665a-config\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.446330 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.466889 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.481111 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/26e1869e-3f67-4f4d-8aad-3007fbdeb4ec-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qkg8t\" (UID: \"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.484500 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.505168 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.514462 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.514615 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/34af5be8-778f-43a8-8241-3ece02457a25-profile-collector-cert\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.514910 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9af915f-4e5b-440d-a3b3-534d16b1f686-secret-volume\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.526020 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.528353 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.528957 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.028935444 +0000 UTC m=+135.679203086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.538130 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-srv-cert\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.546098 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.565548 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.573940 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9af915f-4e5b-440d-a3b3-534d16b1f686-config-volume\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.585556 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.605633 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.624963 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.630207 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.630346 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.130321733 +0000 UTC m=+135.780589365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.630803 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.631301 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.1312845 +0000 UTC m=+135.781552312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.634517 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eddcefaa-1fba-4c03-8b33-886c131532e0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.644574 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.664751 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.673802 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eddcefaa-1fba-4c03-8b33-886c131532e0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.684300 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.692564 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/34af5be8-778f-43a8-8241-3ece02457a25-srv-cert\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.705255 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.714891 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-stats-auth\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.725633 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.732062 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.732206 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.232179773 +0000 UTC m=+135.882447405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.732554 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.733145 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.23312035 +0000 UTC m=+135.883387982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.745219 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.756988 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-default-certificate\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.765318 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.770293 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9ef9416-d92a-401b-8f80-81ab25b24d31-service-ca-bundle\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.785654 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.798376 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9ef9416-d92a-401b-8f80-81ab25b24d31-metrics-certs\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.805612 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.825007 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.835249 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.835486 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.335453534 +0000 UTC m=+135.985721186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.835890 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.836341 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.336320539 +0000 UTC m=+135.986588181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.845170 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.857359 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/297926a5-4372-4868-a1b8-3052078d4217-proxy-tls\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.865346 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.885687 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.895513 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/257d6f9b-ac33-4f18-ab46-2ada42beb644-webhook-cert\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.895926 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/257d6f9b-ac33-4f18-ab46-2ada42beb644-apiservice-cert\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.905056 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.907133 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8498422c-9c5b-46cc-a19b-4c381511abf0-config-volume\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.925436 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.937708 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.938020 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.437979844 +0000 UTC m=+136.088247526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.938606 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:35 crc kubenswrapper[4957]: E1206 05:41:35.939296 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.439270851 +0000 UTC m=+136.089538523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.945932 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.953457 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8498422c-9c5b-46cc-a19b-4c381511abf0-metrics-tls\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.965290 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.972275 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dec5d39a-fa0a-413c-b4a9-014d95bae8e0-cert\") pod \"ingress-canary-nlpbn\" (UID: \"dec5d39a-fa0a-413c-b4a9-014d95bae8e0\") " pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:35 crc kubenswrapper[4957]: I1206 05:41:35.986476 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.004956 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.024994 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.039672 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.039873 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.539809205 +0000 UTC m=+136.190076867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.040074 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.040620 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.540600287 +0000 UTC m=+136.190867959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.074929 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27lcg\" (UniqueName: \"kubernetes.io/projected/b2983a4e-cc8a-4020-986f-920b7732b840-kube-api-access-27lcg\") pod \"apiserver-7bbb656c7d-7qkmz\" (UID: \"b2983a4e-cc8a-4020-986f-920b7732b840\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.091815 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmfwg\" (UniqueName: \"kubernetes.io/projected/74dd4f1f-4ba4-4ee2-9200-c495d70fa447-kube-api-access-bmfwg\") pod \"machine-approver-56656f9798-wsthj\" (UID: \"74dd4f1f-4ba4-4ee2-9200-c495d70fa447\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.111321 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxvn\" (UniqueName: \"kubernetes.io/projected/da2fdc27-1d40-41fa-b257-57b42be7f811-kube-api-access-bnxvn\") pod \"cluster-samples-operator-665b6dd947-djsdh\" (UID: \"da2fdc27-1d40-41fa-b257-57b42be7f811\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.125357 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt2xv\" (UniqueName: \"kubernetes.io/projected/c9105dc1-76e5-46dd-854a-c79f1c000c27-kube-api-access-mt2xv\") pod \"downloads-7954f5f757-ftjwd\" (UID: \"c9105dc1-76e5-46dd-854a-c79f1c000c27\") " pod="openshift-console/downloads-7954f5f757-ftjwd" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.141673 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.641633534 +0000 UTC m=+136.291901196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.141866 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.142728 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.144541 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.644482015 +0000 UTC m=+136.294749687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.162403 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28qm5\" (UniqueName: \"kubernetes.io/projected/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-kube-api-access-28qm5\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.178515 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf99g\" (UniqueName: \"kubernetes.io/projected/70c2a386-5150-4abb-9e97-af87740cc325-kube-api-access-sf99g\") pod \"oauth-openshift-558db77b4-vs6qp\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.184812 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.186570 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-28v9t\" (UID: \"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.193922 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-certs\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.199640 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.206786 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.223088 4957 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.223229 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-node-bootstrap-token podName:574e7efb-63cb-4cbe-a41c-abc7e55d165f nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.723190373 +0000 UTC m=+136.373458035 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-node-bootstrap-token") pod "machine-config-server-5g4vs" (UID: "574e7efb-63cb-4cbe-a41c-abc7e55d165f") : failed to sync secret cache: timed out waiting for the condition Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.226314 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.244348 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.244604 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.744555495 +0000 UTC m=+136.394823127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.245029 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.245580 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.745523232 +0000 UTC m=+136.395790904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.246489 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.265885 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.283378 4957 request.go:700] Waited for 1.873841477s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/secrets?fieldSelector=metadata.name%3Dcsi-hostpath-provisioner-sa-dockercfg-qd74k&limit=500&resourceVersion=0 Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.285050 4957 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.345992 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ftjwd" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.346733 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.347110 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.847059034 +0000 UTC m=+136.497326686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.347480 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.348274 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.848261649 +0000 UTC m=+136.498529291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.353974 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghrg4\" (UniqueName: \"kubernetes.io/projected/2500eb36-b0f0-4dcf-bb78-b18f90b4da34-kube-api-access-ghrg4\") pod \"console-operator-58897d9998-6ktjh\" (UID: \"2500eb36-b0f0-4dcf-bb78-b18f90b4da34\") " pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.372385 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qkw4\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-kube-api-access-6qkw4\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.375647 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.381386 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.389509 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.394868 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-bound-sa-token\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.396528 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.404133 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdpww\" (UniqueName: \"kubernetes.io/projected/a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1-kube-api-access-bdpww\") pod \"authentication-operator-69f744f599-vlwcn\" (UID: \"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.423554 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swrrj\" (UniqueName: \"kubernetes.io/projected/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-kube-api-access-swrrj\") pod \"console-f9d7485db-8mb7l\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:36 crc kubenswrapper[4957]: W1206 05:41:36.433021 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74dd4f1f_4ba4_4ee2_9200_c495d70fa447.slice/crio-8e07f9d18ed7b826b1925e468e2873cfadd644e8fcfb7f4d07c6aa074a48b31e WatchSource:0}: Error finding container 8e07f9d18ed7b826b1925e468e2873cfadd644e8fcfb7f4d07c6aa074a48b31e: Status 404 returned error can't find the container with id 8e07f9d18ed7b826b1925e468e2873cfadd644e8fcfb7f4d07c6aa074a48b31e Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.448886 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.449613 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:36.949592905 +0000 UTC m=+136.599860537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.457794 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz"] Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.468589 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kk7b\" (UniqueName: \"kubernetes.io/projected/0014fb05-9bc2-4ac5-a3b0-ac7abed30219-kube-api-access-7kk7b\") pod \"openshift-config-operator-7777fb866f-284vj\" (UID: \"0014fb05-9bc2-4ac5-a3b0-ac7abed30219\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.469581 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsxnw\" (UniqueName: \"kubernetes.io/projected/b9ad5cbf-adfe-4012-8968-0d7748db277d-kube-api-access-qsxnw\") pod \"machine-api-operator-5694c8668f-c5ld8\" (UID: \"b9ad5cbf-adfe-4012-8968-0d7748db277d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:36 crc kubenswrapper[4957]: W1206 05:41:36.480309 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2983a4e_cc8a_4020_986f_920b7732b840.slice/crio-174d14aca8c91de6ab03316dbd57c557498e4be3528f16f5c198e821dc4b27f3 WatchSource:0}: Error finding container 174d14aca8c91de6ab03316dbd57c557498e4be3528f16f5c198e821dc4b27f3: Status 404 returned error can't find the container with id 174d14aca8c91de6ab03316dbd57c557498e4be3528f16f5c198e821dc4b27f3 Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.491624 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fprz\" (UniqueName: \"kubernetes.io/projected/b9e31a64-6ce7-4b24-9891-9775e7193b8a-kube-api-access-5fprz\") pod \"route-controller-manager-6576b87f9c-vpk8z\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.501518 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vkxr\" (UniqueName: \"kubernetes.io/projected/950d1218-8492-4736-8dbb-cfa12d544bae-kube-api-access-6vkxr\") pod \"controller-manager-879f6c89f-8m4pc\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.513477 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" event={"ID":"74dd4f1f-4ba4-4ee2-9200-c495d70fa447","Type":"ContainerStarted","Data":"8e07f9d18ed7b826b1925e468e2873cfadd644e8fcfb7f4d07c6aa074a48b31e"} Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.514779 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" event={"ID":"b2983a4e-cc8a-4020-986f-920b7732b840","Type":"ContainerStarted","Data":"174d14aca8c91de6ab03316dbd57c557498e4be3528f16f5c198e821dc4b27f3"} Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.530179 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v96m\" (UniqueName: \"kubernetes.io/projected/eddcefaa-1fba-4c03-8b33-886c131532e0-kube-api-access-6v96m\") pod \"openshift-controller-manager-operator-756b6f6bc6-q7zb4\" (UID: \"eddcefaa-1fba-4c03-8b33-886c131532e0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.530683 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.549691 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ps2f\" (UniqueName: \"kubernetes.io/projected/f9af915f-4e5b-440d-a3b3-534d16b1f686-kube-api-access-4ps2f\") pod \"collect-profiles-29416650-9w9j6\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.550627 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.551330 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.051311681 +0000 UTC m=+136.701579313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.551931 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.572170 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.575851 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq8lb\" (UniqueName: \"kubernetes.io/projected/12d0b3b5-412a-42a9-8666-60d47698a3e7-kube-api-access-zq8lb\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.585681 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7665fa45-3b81-48a1-ba82-e6471ae634e9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lthnh\" (UID: \"7665fa45-3b81-48a1-ba82-e6471ae634e9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.592140 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.609077 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jznsp\" (UniqueName: \"kubernetes.io/projected/9a95ad0d-9149-4c33-9f57-a9ca76143509-kube-api-access-jznsp\") pod \"service-ca-operator-777779d784-nz6jh\" (UID: \"9a95ad0d-9149-4c33-9f57-a9ca76143509\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.609276 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.617818 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12d0b3b5-412a-42a9-8666-60d47698a3e7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-k5d5r\" (UID: \"12d0b3b5-412a-42a9-8666-60d47698a3e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.637197 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.652239 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.652711 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zndq5\" (UniqueName: \"kubernetes.io/projected/297926a5-4372-4868-a1b8-3052078d4217-kube-api-access-zndq5\") pod \"machine-config-controller-84d6567774-8kk7m\" (UID: \"297926a5-4372-4868-a1b8-3052078d4217\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.652948 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.152917576 +0000 UTC m=+136.803185208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.660494 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.662930 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w7tl\" (UniqueName: \"kubernetes.io/projected/8498422c-9c5b-46cc-a19b-4c381511abf0-kube-api-access-2w7tl\") pod \"dns-default-tbzz4\" (UID: \"8498422c-9c5b-46cc-a19b-4c381511abf0\") " pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.671392 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ftjwd"] Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.691394 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzd8n\" (UniqueName: \"kubernetes.io/projected/37a32874-c303-4090-b86e-3d2caf34d36a-kube-api-access-tzd8n\") pod \"machine-config-operator-74547568cd-kcqgr\" (UID: \"37a32874-c303-4090-b86e-3d2caf34d36a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.692900 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.706192 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.711276 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t"] Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.711643 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt8lt\" (UniqueName: \"kubernetes.io/projected/cd29b564-939e-4304-9f8e-11aefcb02a5e-kube-api-access-qt8lt\") pod \"apiserver-76f77b778f-hj5wv\" (UID: \"cd29b564-939e-4304-9f8e-11aefcb02a5e\") " pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.717348 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.723410 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.739686 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.739778 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpknx\" (UniqueName: \"kubernetes.io/projected/665360fe-3cbc-49bb-9316-01a6cf62174e-kube-api-access-fpknx\") pod \"marketplace-operator-79b997595-gfv9x\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.746481 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686zt\" (UniqueName: \"kubernetes.io/projected/aa443eb3-221f-42d3-a946-f5bfee0d7148-kube-api-access-686zt\") pod \"csi-hostpathplugin-9zmkv\" (UID: \"aa443eb3-221f-42d3-a946-f5bfee0d7148\") " pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.760646 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.760710 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-node-bootstrap-token\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.764188 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.764934 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtrf8\" (UniqueName: \"kubernetes.io/projected/34af5be8-778f-43a8-8241-3ece02457a25-kube-api-access-jtrf8\") pod \"catalog-operator-68c6474976-k4xws\" (UID: \"34af5be8-778f-43a8-8241-3ece02457a25\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.765104 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.265079217 +0000 UTC m=+136.915347049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.767534 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.777714 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh"] Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.783201 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/574e7efb-63cb-4cbe-a41c-abc7e55d165f-node-bootstrap-token\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.798592 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd4nk\" (UniqueName: \"kubernetes.io/projected/7c4ad024-e034-4584-a801-7f59d97e588f-kube-api-access-gd4nk\") pod \"kube-storage-version-migrator-operator-b67b599dd-7xxjr\" (UID: \"7c4ad024-e034-4584-a801-7f59d97e588f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.806420 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hdb8\" (UniqueName: \"kubernetes.io/projected/26e1869e-3f67-4f4d-8aad-3007fbdeb4ec-kube-api-access-4hdb8\") pod \"multus-admission-controller-857f4d67dd-qkg8t\" (UID: \"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.823315 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sg89\" (UniqueName: \"kubernetes.io/projected/72514156-b413-4c84-802b-f114ebed8090-kube-api-access-4sg89\") pod \"package-server-manager-789f6589d5-9xm5h\" (UID: \"72514156-b413-4c84-802b-f114ebed8090\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.837158 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.844903 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7nrq\" (UniqueName: \"kubernetes.io/projected/574e7efb-63cb-4cbe-a41c-abc7e55d165f-kube-api-access-j7nrq\") pod \"machine-config-server-5g4vs\" (UID: \"574e7efb-63cb-4cbe-a41c-abc7e55d165f\") " pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.854508 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vs6qp"] Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.862194 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.862657 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.362641427 +0000 UTC m=+137.012909059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.863699 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.866823 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqf7m\" (UniqueName: \"kubernetes.io/projected/fa307629-37e4-43eb-bd29-74dda3004967-kube-api-access-fqf7m\") pod \"migrator-59844c95c7-9ckdj\" (UID: \"fa307629-37e4-43eb-bd29-74dda3004967\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.877107 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.891402 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66cmp\" (UniqueName: \"kubernetes.io/projected/92ffc396-7acd-4dd1-8bff-61901a7e6159-kube-api-access-66cmp\") pod \"control-plane-machine-set-operator-78cbb6b69f-m7xn8\" (UID: \"92ffc396-7acd-4dd1-8bff-61901a7e6159\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.895696 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.916473 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vlwcn"] Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.928881 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.934584 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd7ch\" (UniqueName: \"kubernetes.io/projected/dec5d39a-fa0a-413c-b4a9-014d95bae8e0-kube-api-access-fd7ch\") pod \"ingress-canary-nlpbn\" (UID: \"dec5d39a-fa0a-413c-b4a9-014d95bae8e0\") " pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.937625 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.938162 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82xbl\" (UniqueName: \"kubernetes.io/projected/1f2412f9-fb1d-49fa-9337-d9b526ce074d-kube-api-access-82xbl\") pod \"etcd-operator-b45778765-jbd4c\" (UID: \"1f2412f9-fb1d-49fa-9337-d9b526ce074d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.949524 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62sl\" (UniqueName: \"kubernetes.io/projected/257d6f9b-ac33-4f18-ab46-2ada42beb644-kube-api-access-d62sl\") pod \"packageserver-d55dfcdfc-8jbsv\" (UID: \"257d6f9b-ac33-4f18-ab46-2ada42beb644\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.964689 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf7tj\" (UniqueName: \"kubernetes.io/projected/a84a237c-277b-43f3-82ab-6c277059c76c-kube-api-access-tf7tj\") pod \"service-ca-9c57cc56f-qrcnq\" (UID: \"a84a237c-277b-43f3-82ab-6c277059c76c\") " pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.966264 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:36 crc kubenswrapper[4957]: E1206 05:41:36.967514 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.467495873 +0000 UTC m=+137.117763505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.971698 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:36 crc kubenswrapper[4957]: I1206 05:41:36.980985 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhz4k\" (UniqueName: \"kubernetes.io/projected/a6552ccf-41d8-4f61-9ff8-26c3b99c7172-kube-api-access-bhz4k\") pod \"olm-operator-6b444d44fb-4j9pm\" (UID: \"a6552ccf-41d8-4f61-9ff8-26c3b99c7172\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.005801 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.023676 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqnl5\" (UniqueName: \"kubernetes.io/projected/922c58c7-e5f1-4e55-bf25-a903738bea1b-kube-api-access-vqnl5\") pod \"dns-operator-744455d44c-fls7b\" (UID: \"922c58c7-e5f1-4e55-bf25-a903738bea1b\") " pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.028670 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nlpbn" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.038953 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5g4vs" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.046733 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbhmm\" (UniqueName: \"kubernetes.io/projected/f9ef9416-d92a-401b-8f80-81ab25b24d31-kube-api-access-gbhmm\") pod \"router-default-5444994796-9kjdh\" (UID: \"f9ef9416-d92a-401b-8f80-81ab25b24d31\") " pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.049534 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.057417 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.068374 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.068605 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.568566021 +0000 UTC m=+137.218833653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.068798 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.069235 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.56922657 +0000 UTC m=+137.219494202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.080143 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.081250 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.083050 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e299c820-62cc-4e26-90fc-d80670866a0a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c8zkq\" (UID: \"e299c820-62cc-4e26-90fc-d80670866a0a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.108334 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.126406 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.143195 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9zmkv"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.144500 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.155907 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.169426 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.169693 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.669667521 +0000 UTC m=+137.319935163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.169745 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.170109 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.670102673 +0000 UTC m=+137.320370305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.170155 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8mb7l"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.182565 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6ktjh"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.186468 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-284vj"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.187910 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8m4pc"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.203270 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.248183 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.271114 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.271319 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.771295575 +0000 UTC m=+137.421563207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.271604 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.271981 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.771973405 +0000 UTC m=+137.422241037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.280866 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.283363 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c5ld8"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.285773 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z"] Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.373297 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.373681 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.873656611 +0000 UTC m=+137.523924243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.373799 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.374139 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.874121604 +0000 UTC m=+137.524389236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.475041 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.475213 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.975189992 +0000 UTC m=+137.625457624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.475458 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.475883 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:37.97582386 +0000 UTC m=+137.626091502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.577372 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.577527 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.077499537 +0000 UTC m=+137.727767169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.578170 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.578538 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.078526745 +0000 UTC m=+137.728794577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.679346 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.679513 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.179489331 +0000 UTC m=+137.829756963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.679667 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.680007 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.179999305 +0000 UTC m=+137.830266937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.780861 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.781138 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.281090214 +0000 UTC m=+137.931357846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.781530 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.782010 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.281993201 +0000 UTC m=+137.932260833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.890399 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.890882 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.390863538 +0000 UTC m=+138.041131180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:37 crc kubenswrapper[4957]: I1206 05:41:37.991611 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:37 crc kubenswrapper[4957]: E1206 05:41:37.992009 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.49199424 +0000 UTC m=+138.142261862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.092892 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.093295 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.593267144 +0000 UTC m=+138.243534816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.194561 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.195204 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.695173036 +0000 UTC m=+138.345440708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.296099 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.296406 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.796364818 +0000 UTC m=+138.446632500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.296543 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.296939 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.796927454 +0000 UTC m=+138.447195086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.397648 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.397957 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.897928281 +0000 UTC m=+138.548195943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.398138 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.398617 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.8986012 +0000 UTC m=+138.548868862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.499201 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.499807 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:38.999776392 +0000 UTC m=+138.650044074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.601868 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.602500 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.102475327 +0000 UTC m=+138.752742959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.702723 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.703002 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.202961869 +0000 UTC m=+138.853229501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.703216 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.703648 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.203640248 +0000 UTC m=+138.853907880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.790494 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aaeb2a27-18ff-4750-bd39-e24c8e59665a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8stt7\" (UID: \"aaeb2a27-18ff-4750-bd39-e24c8e59665a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.808711 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5lh4\" (UniqueName: \"kubernetes.io/projected/85ba6a69-9f7a-4033-9fcd-25d0ba6f3155-kube-api-access-k5lh4\") pod \"openshift-apiserver-operator-796bbdcf4f-5b8p9\" (UID: \"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.809514 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.809744 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.309698327 +0000 UTC m=+138.959965989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.810154 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.810726 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.310705736 +0000 UTC m=+138.960973368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: W1206 05:41:38.814148 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70c2a386_5150_4abb_9e97_af87740cc325.slice/crio-ebcca660a2754e891cfb8e89967c8a2eebb94238757aae4120009f0af5516bd5 WatchSource:0}: Error finding container ebcca660a2754e891cfb8e89967c8a2eebb94238757aae4120009f0af5516bd5: Status 404 returned error can't find the container with id ebcca660a2754e891cfb8e89967c8a2eebb94238757aae4120009f0af5516bd5 Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.816859 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" event={"ID":"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4","Type":"ContainerStarted","Data":"9ab26cd1cb3b4594f8e78d7be93a457e193f78ecada9c73d2285ff3ca51609ff"} Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.821009 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ftjwd" event={"ID":"c9105dc1-76e5-46dd-854a-c79f1c000c27","Type":"ContainerStarted","Data":"7bb5479522d37a914cc0c9212934093541d9a111bc7170ec6c4acbde3e6e63f5"} Dec 06 05:41:38 crc kubenswrapper[4957]: W1206 05:41:38.831069 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa443eb3_221f_42d3_a946_f5bfee0d7148.slice/crio-63ff7c3b388574e2787cc3bc1d8e57f90d4444023f0fd988b57fe13cb40815dc WatchSource:0}: Error finding container 63ff7c3b388574e2787cc3bc1d8e57f90d4444023f0fd988b57fe13cb40815dc: Status 404 returned error can't find the container with id 63ff7c3b388574e2787cc3bc1d8e57f90d4444023f0fd988b57fe13cb40815dc Dec 06 05:41:38 crc kubenswrapper[4957]: W1206 05:41:38.834103 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0014fb05_9bc2_4ac5_a3b0_ac7abed30219.slice/crio-a8310e9f307a6e0542a094fa6725ac3e5f455edc22f579b4385e414ea7a2e71f WatchSource:0}: Error finding container a8310e9f307a6e0542a094fa6725ac3e5f455edc22f579b4385e414ea7a2e71f: Status 404 returned error can't find the container with id a8310e9f307a6e0542a094fa6725ac3e5f455edc22f579b4385e414ea7a2e71f Dec 06 05:41:38 crc kubenswrapper[4957]: W1206 05:41:38.841580 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda262d0e2_1182_4ba1_92e4_8b0b0c5cf6c1.slice/crio-712a30a0751a8478efe007d70c29eafac7f8e53c3c78535c36811767086e45bb WatchSource:0}: Error finding container 712a30a0751a8478efe007d70c29eafac7f8e53c3c78535c36811767086e45bb: Status 404 returned error can't find the container with id 712a30a0751a8478efe007d70c29eafac7f8e53c3c78535c36811767086e45bb Dec 06 05:41:38 crc kubenswrapper[4957]: W1206 05:41:38.864134 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2500eb36_b0f0_4dcf_bb78_b18f90b4da34.slice/crio-14acb3c2b6dc1a13ecfce2211d887bd15e525ebb4dd9ef0e3c6c962aeb5eb4bd WatchSource:0}: Error finding container 14acb3c2b6dc1a13ecfce2211d887bd15e525ebb4dd9ef0e3c6c962aeb5eb4bd: Status 404 returned error can't find the container with id 14acb3c2b6dc1a13ecfce2211d887bd15e525ebb4dd9ef0e3c6c962aeb5eb4bd Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.928066 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:38 crc kubenswrapper[4957]: E1206 05:41:38.931347 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.431323506 +0000 UTC m=+139.081591138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:38 crc kubenswrapper[4957]: I1206 05:41:38.985219 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.027438 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.030041 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.030401 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.530388928 +0000 UTC m=+139.180656560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.069777 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6"] Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.126267 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh"] Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.131336 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.131489 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.631456776 +0000 UTC m=+139.281724428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.131714 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.132021 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.632010393 +0000 UTC m=+139.282278025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.161816 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hj5wv"] Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.233568 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.234154 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.73411682 +0000 UTC m=+139.384384502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.252392 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gfv9x"] Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.264073 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tbzz4"] Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.329655 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv"] Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.336549 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.336991 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.83697093 +0000 UTC m=+139.487238562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.365764 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qkg8t"] Dec 06 05:41:39 crc kubenswrapper[4957]: W1206 05:41:39.380447 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod665360fe_3cbc_49bb_9316_01a6cf62174e.slice/crio-6d6965ada9c5fe48f7b128fc5fca83a0e9ed747d5d63ee87b8db74d2468c5809 WatchSource:0}: Error finding container 6d6965ada9c5fe48f7b128fc5fca83a0e9ed747d5d63ee87b8db74d2468c5809: Status 404 returned error can't find the container with id 6d6965ada9c5fe48f7b128fc5fca83a0e9ed747d5d63ee87b8db74d2468c5809 Dec 06 05:41:39 crc kubenswrapper[4957]: W1206 05:41:39.394901 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8498422c_9c5b_46cc_a19b_4c381511abf0.slice/crio-82ac8cee7554e9ccab923ab938abb8c56eb79455a5b87e0f488db28368d98586 WatchSource:0}: Error finding container 82ac8cee7554e9ccab923ab938abb8c56eb79455a5b87e0f488db28368d98586: Status 404 returned error can't find the container with id 82ac8cee7554e9ccab923ab938abb8c56eb79455a5b87e0f488db28368d98586 Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.431036 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qrcnq"] Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.438987 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.439680 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:39.939654673 +0000 UTC m=+139.589922305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.540559 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.540990 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.04097652 +0000 UTC m=+139.691244152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.642602 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.643071 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.143056507 +0000 UTC m=+139.793324139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.643422 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.643762 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.143755166 +0000 UTC m=+139.794022798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: W1206 05:41:39.664668 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda84a237c_277b_43f3_82ab_6c277059c76c.slice/crio-f74e4b5992e29fc4e8e79002180fa3d497c338ef9ef5c78a25ba4da48e14f628 WatchSource:0}: Error finding container f74e4b5992e29fc4e8e79002180fa3d497c338ef9ef5c78a25ba4da48e14f628: Status 404 returned error can't find the container with id f74e4b5992e29fc4e8e79002180fa3d497c338ef9ef5c78a25ba4da48e14f628 Dec 06 05:41:39 crc kubenswrapper[4957]: W1206 05:41:39.668515 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9ef9416_d92a_401b_8f80_81ab25b24d31.slice/crio-fdc3217b0296c83b95a1476e38b00ecbd001cfac61083841653b7570d88abb4f WatchSource:0}: Error finding container fdc3217b0296c83b95a1476e38b00ecbd001cfac61083841653b7570d88abb4f: Status 404 returned error can't find the container with id fdc3217b0296c83b95a1476e38b00ecbd001cfac61083841653b7570d88abb4f Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.752340 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.752773 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.252752809 +0000 UTC m=+139.903020441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.851739 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" event={"ID":"b9e31a64-6ce7-4b24-9891-9775e7193b8a","Type":"ContainerStarted","Data":"c721173487db5161059edd627c258646e356fe2a7ca3a6a795787a921a81bc7d"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.852169 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" event={"ID":"b9e31a64-6ce7-4b24-9891-9775e7193b8a","Type":"ContainerStarted","Data":"9add6f18901d248524103582db4541bc2679114fb80880dc60f374f5b92152df"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.853465 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.855437 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.855788 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.355774413 +0000 UTC m=+140.006042045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.855885 4957 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-vpk8z container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.855920 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" podUID="b9e31a64-6ce7-4b24-9891-9775e7193b8a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.876968 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" podStartSLOduration=120.876940339 podStartE2EDuration="2m0.876940339s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:39.876121886 +0000 UTC m=+139.526389538" watchObservedRunningTime="2025-12-06 05:41:39.876940339 +0000 UTC m=+139.527207971" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.880796 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" event={"ID":"f9af915f-4e5b-440d-a3b3-534d16b1f686","Type":"ContainerStarted","Data":"de574bc3179f1507d2559b80e1cf75f068c151c57445cc4a6aa42f7e84051298"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.892067 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" event={"ID":"3dba7d5c-b3c5-4880-a63e-7ac4702eb6d4","Type":"ContainerStarted","Data":"1fd7e1e781b8ddee5699211f82bcd15dad74a56ae46710a4aa1250552cecdfdd"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.897763 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" event={"ID":"b9ad5cbf-adfe-4012-8968-0d7748db277d","Type":"ContainerStarted","Data":"4b7732d23acb2c6b67cee85f82ec2eb995edec153263d2db1a40f411b36d7d1a"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.897847 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" event={"ID":"b9ad5cbf-adfe-4012-8968-0d7748db277d","Type":"ContainerStarted","Data":"e57878dac4fea167dece0b0c5bab334af7501e7f121cbd5087d81828005d90fc"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.919401 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" event={"ID":"74dd4f1f-4ba4-4ee2-9200-c495d70fa447","Type":"ContainerStarted","Data":"2804d1d929162f8365fd97f5d3e1349f4d3cc8aa1f1c080e2cdf61fcec427984"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.929007 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" event={"ID":"7665fa45-3b81-48a1-ba82-e6471ae634e9","Type":"ContainerStarted","Data":"b69f0f3046b0467b76f829838a2f60bc12f668fa5d1d9ded94b8c0a5ccf17440"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.932999 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9kjdh" event={"ID":"f9ef9416-d92a-401b-8f80-81ab25b24d31","Type":"ContainerStarted","Data":"fdc3217b0296c83b95a1476e38b00ecbd001cfac61083841653b7570d88abb4f"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.936377 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" event={"ID":"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec","Type":"ContainerStarted","Data":"aacb4fb8d444bfd5b1eb2fa640971cdd96942a7829ed8d46304c450c69d64ebb"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.955105 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" event={"ID":"950d1218-8492-4736-8dbb-cfa12d544bae","Type":"ContainerStarted","Data":"bfd867228f5dd65e3fbc69a7480136147f047364dc11e4fa4dde4b962cee4468"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.955151 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" event={"ID":"950d1218-8492-4736-8dbb-cfa12d544bae","Type":"ContainerStarted","Data":"8f27967ed207cb88252514667e6835d4e82f7a678949286146db939edb916f4b"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.959264 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.962010 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:39 crc kubenswrapper[4957]: E1206 05:41:39.963122 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.463091177 +0000 UTC m=+140.113358809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.968717 4957 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8m4pc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.968760 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" podUID="950d1218-8492-4736-8dbb-cfa12d544bae" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.972300 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" event={"ID":"2500eb36-b0f0-4dcf-bb78-b18f90b4da34","Type":"ContainerStarted","Data":"0d2675b53d14f118b0db7e32e7f4eac17ab791dffc0e3a6a934042df291093ac"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.972337 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" event={"ID":"2500eb36-b0f0-4dcf-bb78-b18f90b4da34","Type":"ContainerStarted","Data":"14acb3c2b6dc1a13ecfce2211d887bd15e525ebb4dd9ef0e3c6c962aeb5eb4bd"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.973282 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.986041 4957 patch_prober.go:28] interesting pod/console-operator-58897d9998-6ktjh container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.986090 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" podUID="2500eb36-b0f0-4dcf-bb78-b18f90b4da34" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.987075 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" event={"ID":"0014fb05-9bc2-4ac5-a3b0-ac7abed30219","Type":"ContainerStarted","Data":"5590ed1fddc9a2aa41a3468eb335c0993565484f4ff1460f80fb4c90ffea2d41"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.987123 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" event={"ID":"0014fb05-9bc2-4ac5-a3b0-ac7abed30219","Type":"ContainerStarted","Data":"a8310e9f307a6e0542a094fa6725ac3e5f455edc22f579b4385e414ea7a2e71f"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.990216 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-28v9t" podStartSLOduration=120.990199432 podStartE2EDuration="2m0.990199432s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:39.922528994 +0000 UTC m=+139.572796646" watchObservedRunningTime="2025-12-06 05:41:39.990199432 +0000 UTC m=+139.640467064" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.992944 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" podStartSLOduration=120.992934079 podStartE2EDuration="2m0.992934079s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:39.98839525 +0000 UTC m=+139.638662892" watchObservedRunningTime="2025-12-06 05:41:39.992934079 +0000 UTC m=+139.643201711" Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.993611 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" event={"ID":"cd29b564-939e-4304-9f8e-11aefcb02a5e","Type":"ContainerStarted","Data":"4ec4e64811d2690f601cfa655ddf73c7278d6bce5116acf4d0965c7b45892ef4"} Dec 06 05:41:39 crc kubenswrapper[4957]: I1206 05:41:39.999721 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ftjwd" event={"ID":"c9105dc1-76e5-46dd-854a-c79f1c000c27","Type":"ContainerStarted","Data":"f384a19a1ced7670f8cf2de2de24eff5d920ddf2a0c746580073cfc55c93ee33"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.002092 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ftjwd" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.006040 4957 generic.go:334] "Generic (PLEG): container finished" podID="b2983a4e-cc8a-4020-986f-920b7732b840" containerID="d03b789c3ae0db874cca059844cd24eb7247727233997838635665bbac575908" exitCode=0 Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.006133 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" event={"ID":"b2983a4e-cc8a-4020-986f-920b7732b840","Type":"ContainerDied","Data":"d03b789c3ae0db874cca059844cd24eb7247727233997838635665bbac575908"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.012109 4957 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftjwd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.012158 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ftjwd" podUID="c9105dc1-76e5-46dd-854a-c79f1c000c27" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.023346 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8mb7l" event={"ID":"0483ff2c-b2e8-44f2-bec5-85eb4d06d279","Type":"ContainerStarted","Data":"810ab1c8eead3866218d55cf51d0eb564320b73776026fdd3bc9cf753fe1b3e4"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.023393 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8mb7l" event={"ID":"0483ff2c-b2e8-44f2-bec5-85eb4d06d279","Type":"ContainerStarted","Data":"a6fc479e6daf88890e0bc7d5738329117bf7a04d752a649ab872dd5f47e2ac9c"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.042700 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" podStartSLOduration=121.042679681 podStartE2EDuration="2m1.042679681s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:40.037792603 +0000 UTC m=+139.688060245" watchObservedRunningTime="2025-12-06 05:41:40.042679681 +0000 UTC m=+139.692947313" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.063017 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ftjwd" podStartSLOduration=121.062998784 podStartE2EDuration="2m1.062998784s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:40.057647133 +0000 UTC m=+139.707914775" watchObservedRunningTime="2025-12-06 05:41:40.062998784 +0000 UTC m=+139.713266416" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.063709 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.065925 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.565914036 +0000 UTC m=+140.216181668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.080452 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8mb7l" podStartSLOduration=121.080430575 podStartE2EDuration="2m1.080430575s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:40.077021019 +0000 UTC m=+139.727288661" watchObservedRunningTime="2025-12-06 05:41:40.080430575 +0000 UTC m=+139.730698217" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.094375 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" event={"ID":"eddcefaa-1fba-4c03-8b33-886c131532e0","Type":"ContainerStarted","Data":"cfa8b80b98694b4447cb8cb9d26a330734893fcbab9aafcd83dd0fe4a99f3868"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.094450 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" event={"ID":"eddcefaa-1fba-4c03-8b33-886c131532e0","Type":"ContainerStarted","Data":"270fe9ead383475c1b5e5b6ce94c90534d23b5dd917d34ffe1a1154d51c6b643"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.102815 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" event={"ID":"70c2a386-5150-4abb-9e97-af87740cc325","Type":"ContainerStarted","Data":"ebcca660a2754e891cfb8e89967c8a2eebb94238757aae4120009f0af5516bd5"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.121758 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q7zb4" podStartSLOduration=121.121739069 podStartE2EDuration="2m1.121739069s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:40.118457227 +0000 UTC m=+139.768724869" watchObservedRunningTime="2025-12-06 05:41:40.121739069 +0000 UTC m=+139.772006701" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.135009 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" event={"ID":"aa443eb3-221f-42d3-a946-f5bfee0d7148","Type":"ContainerStarted","Data":"63ff7c3b388574e2787cc3bc1d8e57f90d4444023f0fd988b57fe13cb40815dc"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.138052 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" event={"ID":"257d6f9b-ac33-4f18-ab46-2ada42beb644","Type":"ContainerStarted","Data":"781dc204c1d9a3f165d8f3ef67930f78392c192e67e01a6bd91d1d84c462c65d"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.142487 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" event={"ID":"297926a5-4372-4868-a1b8-3052078d4217","Type":"ContainerStarted","Data":"04145449fa93993f82d29e2e8476af97524d0ba0f349a573b555af6780166b4f"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.142561 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" event={"ID":"297926a5-4372-4868-a1b8-3052078d4217","Type":"ContainerStarted","Data":"4546b27934d7a13cf6c73508f72156321d9403481e4103c03eaa3bb542180716"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.152245 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5g4vs" event={"ID":"574e7efb-63cb-4cbe-a41c-abc7e55d165f","Type":"ContainerStarted","Data":"168becd83b9af4201fb0ad49ffc9835cfc8c304c152c87ba7cfaa38fc9b8684e"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.164874 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.165751 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tbzz4" event={"ID":"8498422c-9c5b-46cc-a19b-4c381511abf0","Type":"ContainerStarted","Data":"82ac8cee7554e9ccab923ab938abb8c56eb79455a5b87e0f488db28368d98586"} Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.165871 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.665852243 +0000 UTC m=+140.316119875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.185977 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.186035 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.211280 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" event={"ID":"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1","Type":"ContainerStarted","Data":"1b7a54dae78d311b6c72eac74ee7c80553c754a68449116e0136679eacb9a51c"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.212312 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" event={"ID":"a262d0e2-1182-4ba1-92e4-8b0b0c5cf6c1","Type":"ContainerStarted","Data":"712a30a0751a8478efe007d70c29eafac7f8e53c3c78535c36811767086e45bb"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.217626 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" event={"ID":"a84a237c-277b-43f3-82ab-6c277059c76c","Type":"ContainerStarted","Data":"f74e4b5992e29fc4e8e79002180fa3d497c338ef9ef5c78a25ba4da48e14f628"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.228820 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" event={"ID":"665360fe-3cbc-49bb-9316-01a6cf62174e","Type":"ContainerStarted","Data":"6d6965ada9c5fe48f7b128fc5fca83a0e9ed747d5d63ee87b8db74d2468c5809"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.238871 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-vlwcn" podStartSLOduration=121.238854911 podStartE2EDuration="2m1.238854911s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:40.236232147 +0000 UTC m=+139.886499809" watchObservedRunningTime="2025-12-06 05:41:40.238854911 +0000 UTC m=+139.889122543" Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.256590 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" event={"ID":"da2fdc27-1d40-41fa-b257-57b42be7f811","Type":"ContainerStarted","Data":"0211c15708f80c293a0540d64d2c620ae52df086ae03225cd29db543dc61408a"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.256674 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" event={"ID":"da2fdc27-1d40-41fa-b257-57b42be7f811","Type":"ContainerStarted","Data":"76787c169b2dfeb3e3b02e533e70c652857c6dde9de6ff9e7b997c30fd73f909"} Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.266486 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.271958 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.771823949 +0000 UTC m=+140.422091791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.370344 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.371935 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.871822408 +0000 UTC m=+140.522090060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.472468 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.472812 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:40.972800954 +0000 UTC m=+140.623068586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.573703 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.574123 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.074101629 +0000 UTC m=+140.724369261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.674965 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.685297 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.185266253 +0000 UTC m=+140.835533885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.805247 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.805717 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.305698857 +0000 UTC m=+140.955966489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.824100 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r"] Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.826803 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq"] Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.908245 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:40 crc kubenswrapper[4957]: E1206 05:41:40.910067 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.410040178 +0000 UTC m=+141.060307810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:40 crc kubenswrapper[4957]: I1206 05:41:40.934021 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.013251 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.013581 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.513566236 +0000 UTC m=+141.163833868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.016005 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.033275 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nlpbn"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.036866 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.042560 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.068554 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.077285 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jbd4c"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.114634 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.115322 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.615310535 +0000 UTC m=+141.265578167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.123750 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.125496 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.153613 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.170653 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fls7b"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.195324 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr"] Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.215949 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.216130 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9"] Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.216224 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.716181408 +0000 UTC m=+141.366449030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.216301 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.217252 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.717226857 +0000 UTC m=+141.367494489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.264544 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" event={"ID":"12d0b3b5-412a-42a9-8666-60d47698a3e7","Type":"ContainerStarted","Data":"b39407ea7f5ace512607a9b0f6e051834d9e1c86f84a8a8d811a2a05737a11fb"} Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.266405 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" event={"ID":"e299c820-62cc-4e26-90fc-d80670866a0a","Type":"ContainerStarted","Data":"72903803d34b089acb6c03f82e51e227480cc3b171cedb430d9fa32fa016e996"} Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.269249 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" event={"ID":"a6552ccf-41d8-4f61-9ff8-26c3b99c7172","Type":"ContainerStarted","Data":"403e0205cc00bd7f1244dfa89f5b185f763338800d4a9bdc758ce4788cfbb528"} Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.270857 4957 generic.go:334] "Generic (PLEG): container finished" podID="0014fb05-9bc2-4ac5-a3b0-ac7abed30219" containerID="5590ed1fddc9a2aa41a3468eb335c0993565484f4ff1460f80fb4c90ffea2d41" exitCode=0 Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.270919 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" event={"ID":"0014fb05-9bc2-4ac5-a3b0-ac7abed30219","Type":"ContainerDied","Data":"5590ed1fddc9a2aa41a3468eb335c0993565484f4ff1460f80fb4c90ffea2d41"} Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.272658 4957 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftjwd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.272701 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ftjwd" podUID="c9105dc1-76e5-46dd-854a-c79f1c000c27" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.272016 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" event={"ID":"72514156-b413-4c84-802b-f114ebed8090","Type":"ContainerStarted","Data":"0b575bdc94592d6525c3270824c177d9e98e572fcfd7820214cd109aecc0f89d"} Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.279588 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.280806 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.333309 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.337281 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.83724282 +0000 UTC m=+141.487510452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.337882 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.341151 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.841125459 +0000 UTC m=+141.491393091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.427351 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-6ktjh" Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.454776 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.455221 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:41.955198835 +0000 UTC m=+141.605466467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.556075 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.556664 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.056652784 +0000 UTC m=+141.706920416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.661952 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.662165 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.162138568 +0000 UTC m=+141.812406200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.663151 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.663641 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.163615649 +0000 UTC m=+141.813883281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.764041 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.764382 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.264332758 +0000 UTC m=+141.914600390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.764611 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.765068 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.265047218 +0000 UTC m=+141.915314850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.865504 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.865826 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.365743346 +0000 UTC m=+142.016010978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.865961 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.866345 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.366334893 +0000 UTC m=+142.016602525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:41 crc kubenswrapper[4957]: I1206 05:41:41.966765 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:41 crc kubenswrapper[4957]: E1206 05:41:41.967119 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.467103814 +0000 UTC m=+142.117371446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.071116 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.071576 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.571558098 +0000 UTC m=+142.221825730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.173398 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.173916 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.673900972 +0000 UTC m=+142.324168604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.275800 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.276597 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.776555686 +0000 UTC m=+142.426823318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.388015 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.390783 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.890751975 +0000 UTC m=+142.541019607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.494740 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.495088 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:42.995074995 +0000 UTC m=+142.645342627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.497619 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9kjdh" event={"ID":"f9ef9416-d92a-401b-8f80-81ab25b24d31","Type":"ContainerStarted","Data":"cfed6c14739247ed091c8051e190b614efbbcc43d17f8aeb923c3bb8c4066b82"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.547899 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9kjdh" podStartSLOduration=123.547881473 podStartE2EDuration="2m3.547881473s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:42.546758221 +0000 UTC m=+142.197025863" watchObservedRunningTime="2025-12-06 05:41:42.547881473 +0000 UTC m=+142.198149105" Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.550557 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" event={"ID":"37a32874-c303-4090-b86e-3d2caf34d36a","Type":"ContainerStarted","Data":"06606708d39308f21b4f7411377a4b835ee29a475ffb14afd375d932b73c527b"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.582214 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" event={"ID":"92ffc396-7acd-4dd1-8bff-61901a7e6159","Type":"ContainerStarted","Data":"a20f83cc548d703ca27a35248618e3a79153e3c5f2199eca5c66bd23caea6e3c"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.614848 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.618344 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.118306858 +0000 UTC m=+142.768574490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.629429 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" event={"ID":"fa307629-37e4-43eb-bd29-74dda3004967","Type":"ContainerStarted","Data":"6394454e451d6153f2205b075b577a0088cc9c92b7ad85049ca97e17c68597ec"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.659411 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" event={"ID":"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155","Type":"ContainerStarted","Data":"92c9aa19e0637f9c63dfc5ccfa9483f4aeaab26063a9b9e5c59c9aee8d5c6a3f"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.718971 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.719442 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.219424828 +0000 UTC m=+142.869692460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.747808 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" event={"ID":"922c58c7-e5f1-4e55-bf25-a903738bea1b","Type":"ContainerStarted","Data":"d916f905c93313aaefda30b72d52428e57afac4b549c49f9d753158a59a556f7"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.756370 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" event={"ID":"b2983a4e-cc8a-4020-986f-920b7732b840","Type":"ContainerStarted","Data":"efb61cb70f3be916ce0f21a77bbf81fb3c76163e3ce3706df5e0b065aad5d5ed"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.820084 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.820608 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" event={"ID":"34af5be8-778f-43a8-8241-3ece02457a25","Type":"ContainerStarted","Data":"813537e77368c3aa3bef05bd68e05c905949264cd10b40e48ffaa964c430a660"} Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.821212 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.321173366 +0000 UTC m=+142.971440988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.896716 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.918023 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" event={"ID":"297926a5-4372-4868-a1b8-3052078d4217","Type":"ContainerStarted","Data":"0b6e762d2b566f619797644d66622d15edff3421042125e914833d48be197dc8"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.922566 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:42 crc kubenswrapper[4957]: E1206 05:41:42.923176 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.423153461 +0000 UTC m=+143.073421093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.941987 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" podStartSLOduration=123.941949311 podStartE2EDuration="2m3.941949311s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:42.818054458 +0000 UTC m=+142.468322100" watchObservedRunningTime="2025-12-06 05:41:42.941949311 +0000 UTC m=+142.592216953" Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.942488 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" podStartSLOduration=124.942482436 podStartE2EDuration="2m4.942482436s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:42.939296076 +0000 UTC m=+142.589563728" watchObservedRunningTime="2025-12-06 05:41:42.942482436 +0000 UTC m=+142.592750068" Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.965197 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" event={"ID":"7c4ad024-e034-4584-a801-7f59d97e588f","Type":"ContainerStarted","Data":"c2a874c947d68ad7102e796cbda9fa535d1490d7499c0dbb87dc5c9b902b297b"} Dec 06 05:41:42 crc kubenswrapper[4957]: I1206 05:41:42.973463 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" event={"ID":"9a95ad0d-9149-4c33-9f57-a9ca76143509","Type":"ContainerStarted","Data":"2310df42e05762c062c1aceb192d0cd2652d9169484ee9910369feb653b31bb9"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.005482 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.006680 4957 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vs6qp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" start-of-body= Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.006737 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" podUID="70c2a386-5150-4abb-9e97-af87740cc325" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.018238 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" event={"ID":"aaeb2a27-18ff-4750-bd39-e24c8e59665a","Type":"ContainerStarted","Data":"20ec8fe96b04520ac8f866c12431d97eb6dd5eb2d87f41cfb0d23c3827b75fd1"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.020754 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8kk7m" podStartSLOduration=124.020742312 podStartE2EDuration="2m4.020742312s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.017912902 +0000 UTC m=+142.668180534" watchObservedRunningTime="2025-12-06 05:41:43.020742312 +0000 UTC m=+142.671009944" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.027370 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.027507 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.527486212 +0000 UTC m=+143.177753844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.028322 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.030808 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.530787995 +0000 UTC m=+143.181055627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.042310 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nlpbn" event={"ID":"dec5d39a-fa0a-413c-b4a9-014d95bae8e0","Type":"ContainerStarted","Data":"06cbed5f57c9347f9c216457fed64a34315785a20d8554c08770cec8bb19e8ae"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.061911 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" podStartSLOduration=124.061884971 podStartE2EDuration="2m4.061884971s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.056361175 +0000 UTC m=+142.706628807" watchObservedRunningTime="2025-12-06 05:41:43.061884971 +0000 UTC m=+142.712152593" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.072690 4957 generic.go:334] "Generic (PLEG): container finished" podID="cd29b564-939e-4304-9f8e-11aefcb02a5e" containerID="7d1c67db2088d2bf617c3930c5beeb5addade0d3e72271935fbe4453be085dda" exitCode=0 Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.072865 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" event={"ID":"cd29b564-939e-4304-9f8e-11aefcb02a5e","Type":"ContainerDied","Data":"7d1c67db2088d2bf617c3930c5beeb5addade0d3e72271935fbe4453be085dda"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.125248 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" podStartSLOduration=124.125218417 podStartE2EDuration="2m4.125218417s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.124952889 +0000 UTC m=+142.775220521" watchObservedRunningTime="2025-12-06 05:41:43.125218417 +0000 UTC m=+142.775486039" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.135866 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.136080 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.636036492 +0000 UTC m=+143.286304114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.136354 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.140296 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.64026919 +0000 UTC m=+143.290536822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.230267 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.244095 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.244610 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.744582151 +0000 UTC m=+143.394849783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.249977 4957 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-4j9pm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.250037 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" podUID="a6552ccf-41d8-4f61-9ff8-26c3b99c7172" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.294014 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nlpbn" podStartSLOduration=9.293999214 podStartE2EDuration="9.293999214s" podCreationTimestamp="2025-12-06 05:41:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.2165177 +0000 UTC m=+142.866785332" watchObservedRunningTime="2025-12-06 05:41:43.293999214 +0000 UTC m=+142.944266846" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.297583 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.310345 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:43 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:43 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:43 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.310394 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.355272 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.355659 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.855642961 +0000 UTC m=+143.505910603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.362181 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" event={"ID":"1f2412f9-fb1d-49fa-9337-d9b526ce074d","Type":"ContainerStarted","Data":"f889693ba2e7d7f1f95d590801f6264a49de9d35916e3495e5962ea2af59a349"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.456731 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.460356 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:43.960328842 +0000 UTC m=+143.610596474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.482794 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" podStartSLOduration=124.482779275 podStartE2EDuration="2m4.482779275s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.422607599 +0000 UTC m=+143.072875231" watchObservedRunningTime="2025-12-06 05:41:43.482779275 +0000 UTC m=+143.133046907" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.493752 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" event={"ID":"aa443eb3-221f-42d3-a946-f5bfee0d7148","Type":"ContainerStarted","Data":"f6af46e659139634aa00b40fcfd256e92a19f1ca42e73a66fd62fb5284eab356"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.515575 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5g4vs" event={"ID":"574e7efb-63cb-4cbe-a41c-abc7e55d165f","Type":"ContainerStarted","Data":"a67101ac49ea98844911112c48334acc8c21a71dea1fec92e48b58ebe9a1ad45"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.560433 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.561977 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.061960367 +0000 UTC m=+143.712227999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.572571 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.585368 4957 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gfv9x container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.585422 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" podUID="665360fe-3cbc-49bb-9316-01a6cf62174e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.603651 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" podStartSLOduration=125.603628311 podStartE2EDuration="2m5.603628311s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.48258573 +0000 UTC m=+143.132853362" watchObservedRunningTime="2025-12-06 05:41:43.603628311 +0000 UTC m=+143.253895943" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.603783 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5g4vs" podStartSLOduration=9.603778355 podStartE2EDuration="9.603778355s" podCreationTimestamp="2025-12-06 05:41:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.577049072 +0000 UTC m=+143.227316714" watchObservedRunningTime="2025-12-06 05:41:43.603778355 +0000 UTC m=+143.254045987" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.663609 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" podStartSLOduration=124.66357029 podStartE2EDuration="2m4.66357029s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.639961955 +0000 UTC m=+143.290229607" watchObservedRunningTime="2025-12-06 05:41:43.66357029 +0000 UTC m=+143.313837922" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.665553 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.665739 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.165722951 +0000 UTC m=+143.815990583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.666585 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.677398 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.177362329 +0000 UTC m=+143.827629951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.731916 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" podStartSLOduration=124.731896037 podStartE2EDuration="2m4.731896037s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.681042493 +0000 UTC m=+143.331310135" watchObservedRunningTime="2025-12-06 05:41:43.731896037 +0000 UTC m=+143.382163669" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.750565 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" event={"ID":"e299c820-62cc-4e26-90fc-d80670866a0a","Type":"ContainerStarted","Data":"6cfce46e7dba6f3c5c5b8b76085280bdc5f662b47ef7fc7847baa9d587254428"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.774543 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" podStartSLOduration=125.774528788 podStartE2EDuration="2m5.774528788s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.73306731 +0000 UTC m=+143.383334952" watchObservedRunningTime="2025-12-06 05:41:43.774528788 +0000 UTC m=+143.424796420" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.780324 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.781596 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.281580807 +0000 UTC m=+143.931848439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.788779 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" event={"ID":"7665fa45-3b81-48a1-ba82-e6471ae634e9","Type":"ContainerStarted","Data":"b4e530ee9fc8934ba3edbfc5b8efa4acb6fff63e236801bd825dffe681b7da61"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.830139 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" podStartSLOduration=124.830119935 podStartE2EDuration="2m4.830119935s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.775462364 +0000 UTC m=+143.425729996" watchObservedRunningTime="2025-12-06 05:41:43.830119935 +0000 UTC m=+143.480387567" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.830234 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" event={"ID":"74dd4f1f-4ba4-4ee2-9200-c495d70fa447","Type":"ContainerStarted","Data":"42d932b42c7f4839be4c9d88eeb233b4f3041da29c2fe3486b6aacedab66db67"} Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.831395 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c8zkq" podStartSLOduration=124.831390351 podStartE2EDuration="2m4.831390351s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.830897007 +0000 UTC m=+143.481164639" watchObservedRunningTime="2025-12-06 05:41:43.831390351 +0000 UTC m=+143.481657983" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.863890 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lthnh" podStartSLOduration=124.863872907 podStartE2EDuration="2m4.863872907s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.862348383 +0000 UTC m=+143.512616025" watchObservedRunningTime="2025-12-06 05:41:43.863872907 +0000 UTC m=+143.514140529" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.884542 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.885904 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.385891478 +0000 UTC m=+144.036159110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.900189 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wsthj" podStartSLOduration=125.90017012 podStartE2EDuration="2m5.90017012s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:43.898478402 +0000 UTC m=+143.548746034" watchObservedRunningTime="2025-12-06 05:41:43.90017012 +0000 UTC m=+143.550437752" Dec 06 05:41:43 crc kubenswrapper[4957]: I1206 05:41:43.985258 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:43 crc kubenswrapper[4957]: E1206 05:41:43.985990 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.485973518 +0000 UTC m=+144.136241160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.087537 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.087874 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.58785847 +0000 UTC m=+144.238126102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.188807 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.189483 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.689461954 +0000 UTC m=+144.339729586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.283646 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:44 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:44 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:44 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.283745 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.290708 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.291193 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.79116719 +0000 UTC m=+144.441434882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.391641 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.391812 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.891783996 +0000 UTC m=+144.542051628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.392058 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.392407 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.892398093 +0000 UTC m=+144.542665725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.492805 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.492992 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.992968178 +0000 UTC m=+144.643235810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.493457 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.493811 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:44.993803362 +0000 UTC m=+144.644070984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.595086 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.595284 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.095258882 +0000 UTC m=+144.745526514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.595410 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.595719 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.095712495 +0000 UTC m=+144.745980127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.695977 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.696326 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.19630994 +0000 UTC m=+144.846577572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.797393 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.797696 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.297685867 +0000 UTC m=+144.947953499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.834634 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" event={"ID":"257d6f9b-ac33-4f18-ab46-2ada42beb644","Type":"ContainerStarted","Data":"747b9d11e72269c6441a431a84cd9bbb33bf699aeefd037b8df19a1c1fce5a8c"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.835504 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.836907 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" event={"ID":"a6552ccf-41d8-4f61-9ff8-26c3b99c7172","Type":"ContainerStarted","Data":"726bf2cd5acba1e2a522b8af133fc72e7cae003a845f832f940a267d27048c12"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.838368 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qrcnq" event={"ID":"a84a237c-277b-43f3-82ab-6c277059c76c","Type":"ContainerStarted","Data":"8dfb180391e7cbdd9f06d0423c3bd049b2eeedac013fe59bc953ff7a7f14eb04"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.839672 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" event={"ID":"1f2412f9-fb1d-49fa-9337-d9b526ce074d","Type":"ContainerStarted","Data":"1b016b72a5ae7a5f5cd128a49d46f0eba6bbc8d1a06417a452afde2b3928ef5d"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.841157 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" event={"ID":"7c4ad024-e034-4584-a801-7f59d97e588f","Type":"ContainerStarted","Data":"99e0511bbaa871c4e794b5501b467c22e9ff565e4775c009ebc7e3af6cdf9c94"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.841311 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9pm" Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.842486 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" event={"ID":"85ba6a69-9f7a-4033-9fcd-25d0ba6f3155","Type":"ContainerStarted","Data":"0b42dcf0e78f0aa375326524801ec295142eb77c3cb106c1259af0fc2e8d612e"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.844140 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" event={"ID":"f9af915f-4e5b-440d-a3b3-534d16b1f686","Type":"ContainerStarted","Data":"cef6bc9892b207b2f3b961a1b79ffebf77071515014aeeb07dbe38b4fc6a239c"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.845165 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" event={"ID":"92ffc396-7acd-4dd1-8bff-61901a7e6159","Type":"ContainerStarted","Data":"7e78ef7c2ee93fb9675fd798779d14342fae6913c5f77207e81c2ed153be0a9c"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.846369 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" event={"ID":"9a95ad0d-9149-4c33-9f57-a9ca76143509","Type":"ContainerStarted","Data":"52b00a1c4b3d009d4cc08c1883e701d3bc3cf0ae69c89b220a49bff83fc01607"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.847717 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5ld8" event={"ID":"b9ad5cbf-adfe-4012-8968-0d7748db277d","Type":"ContainerStarted","Data":"9c7e184c60b69ad5ddfe909f4b20244a67180efaa191168208f073be79184393"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.849008 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" event={"ID":"70c2a386-5150-4abb-9e97-af87740cc325","Type":"ContainerStarted","Data":"90640a5a65d1169b1443983bce4931a15c506353e8de3c5f88b05aef67b07cb4"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.851015 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" event={"ID":"0014fb05-9bc2-4ac5-a3b0-ac7abed30219","Type":"ContainerStarted","Data":"374af5da6bd20db828a6007e9aa81508828c48d04b480b7c13345de00ee96055"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.852237 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nlpbn" event={"ID":"dec5d39a-fa0a-413c-b4a9-014d95bae8e0","Type":"ContainerStarted","Data":"4dc15e4e7c285a1b27f112c789b21e045b19bf33a7d483422ca0659b0e5f0685"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.854014 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" event={"ID":"72514156-b413-4c84-802b-f114ebed8090","Type":"ContainerStarted","Data":"2691831ee164166cf8ed91cf38235df6731b2de44675453b4e9cedbd5e195a67"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.854034 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" event={"ID":"72514156-b413-4c84-802b-f114ebed8090","Type":"ContainerStarted","Data":"400fb88d46114a5e78ce1718f6cd38a69794a80253a18842ed8b6ff30443c223"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.854349 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.856503 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" event={"ID":"cd29b564-939e-4304-9f8e-11aefcb02a5e","Type":"ContainerStarted","Data":"a9e0688ec3fe240aaedda59a016603d5fbe3a01fa86d5caf48e5f0e97cfac474"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.858110 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tbzz4" event={"ID":"8498422c-9c5b-46cc-a19b-4c381511abf0","Type":"ContainerStarted","Data":"81183bc4471702b6382e11678e4796ab9dbbc2db7d8e16e0f58aa49d06f41962"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.858158 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tbzz4" event={"ID":"8498422c-9c5b-46cc-a19b-4c381511abf0","Type":"ContainerStarted","Data":"754c3b58db337347349df2080a3b7b40b36cdb3c82c3f83fe2ffba9b96435f3b"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.858213 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.859768 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" event={"ID":"922c58c7-e5f1-4e55-bf25-a903738bea1b","Type":"ContainerStarted","Data":"644cb2bcdbd80ae2dd970597d4be738595358af8af4b48d7758af5e5c39f9e70"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.859795 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" event={"ID":"922c58c7-e5f1-4e55-bf25-a903738bea1b","Type":"ContainerStarted","Data":"c60738336ea1eb6d78ba34351567ddce587529eca8101a443c5e557a47f61775"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.860977 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" event={"ID":"665360fe-3cbc-49bb-9316-01a6cf62174e","Type":"ContainerStarted","Data":"c71168b6de9b9d702c8a039537e6c5862ef60811a4ea53f8d23db5ef00314d83"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.861413 4957 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gfv9x container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.861443 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" podUID="665360fe-3cbc-49bb-9316-01a6cf62174e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.862321 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" event={"ID":"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec","Type":"ContainerStarted","Data":"ece75fa43afbb4e694ae516a0c2ab0c668b964094e3bef3eb995732bf64a2930"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.862347 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" event={"ID":"26e1869e-3f67-4f4d-8aad-3007fbdeb4ec","Type":"ContainerStarted","Data":"985b637f661f549456b51c1f28cd43eab5732d1d7ca04700110480cb7748381a"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.863849 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" event={"ID":"12d0b3b5-412a-42a9-8666-60d47698a3e7","Type":"ContainerStarted","Data":"e200be8def025f41f9650c8bdaf2695e8182932785121563d500464ce80805a7"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.863876 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-k5d5r" event={"ID":"12d0b3b5-412a-42a9-8666-60d47698a3e7","Type":"ContainerStarted","Data":"7621f50f53c8c37dcf305b41721d585160ca2f9183a27786aa31319ef2ab8689"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.865056 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" event={"ID":"aaeb2a27-18ff-4750-bd39-e24c8e59665a","Type":"ContainerStarted","Data":"7797584fb11e4a4a5cc8f8581bd3b57662c4bf4adfb6e75392e6b0a6752d77f8"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.866274 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" event={"ID":"34af5be8-778f-43a8-8241-3ece02457a25","Type":"ContainerStarted","Data":"f884ec0b05a519835f30ede7f1ce3a6ecdd7df146947921872eaef22e7c6c53b"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.866452 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.867916 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" event={"ID":"fa307629-37e4-43eb-bd29-74dda3004967","Type":"ContainerStarted","Data":"848ec99739fb9a34f36c272682502150a834bbb45c4f5c50bae4d16f94b51f3e"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.867937 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" event={"ID":"fa307629-37e4-43eb-bd29-74dda3004967","Type":"ContainerStarted","Data":"be528b855abefdeeddc59710a2f88f425afa78c3f34686793d76ee68716f6155"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.869338 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-djsdh" event={"ID":"da2fdc27-1d40-41fa-b257-57b42be7f811","Type":"ContainerStarted","Data":"56d161ca89e23823fc62052bb139d2713fd26db54149d0652dfdaceb01f7bf10"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.872579 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" event={"ID":"37a32874-c303-4090-b86e-3d2caf34d36a","Type":"ContainerStarted","Data":"efa5c8ca17507e5213de0e3380f00656829bec7a38e2b73508c7c048d5775026"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.872623 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" event={"ID":"37a32874-c303-4090-b86e-3d2caf34d36a","Type":"ContainerStarted","Data":"c812decb7407fe1d52a90eeb8d12b535faca0830ba8b2228955991c151933685"} Dec 06 05:41:44 crc kubenswrapper[4957]: I1206 05:41:44.898209 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:44 crc kubenswrapper[4957]: E1206 05:41:44.900960 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.400930358 +0000 UTC m=+145.051198000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.000883 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.001379 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.501358878 +0000 UTC m=+145.151626570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.044310 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.074344 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.087924 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.102720 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.103159 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.603134047 +0000 UTC m=+145.253401679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.156988 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8jbsv" podStartSLOduration=126.156954344 podStartE2EDuration="2m6.156954344s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.12168791 +0000 UTC m=+144.771955552" watchObservedRunningTime="2025-12-06 05:41:45.156954344 +0000 UTC m=+144.807221966" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.159071 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kcqgr" podStartSLOduration=126.159065403 podStartE2EDuration="2m6.159065403s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.147290922 +0000 UTC m=+144.797558564" watchObservedRunningTime="2025-12-06 05:41:45.159065403 +0000 UTC m=+144.809333035" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.208363 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.208947 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.708916699 +0000 UTC m=+145.359184511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.271793 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-tbzz4" podStartSLOduration=11.27177679 podStartE2EDuration="11.27177679s" podCreationTimestamp="2025-12-06 05:41:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.271400079 +0000 UTC m=+144.921667711" watchObservedRunningTime="2025-12-06 05:41:45.27177679 +0000 UTC m=+144.922044422" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.291745 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:45 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:45 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:45 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.291936 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.310305 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.310802 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.810749719 +0000 UTC m=+145.461017351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.311112 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.311601 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.811591162 +0000 UTC m=+145.461858794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.339402 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7xxjr" podStartSLOduration=126.339378695 podStartE2EDuration="2m6.339378695s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.337647777 +0000 UTC m=+144.987915409" watchObservedRunningTime="2025-12-06 05:41:45.339378695 +0000 UTC m=+144.989646327" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.414265 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.414654 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:45.914633347 +0000 UTC m=+145.564900979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.467773 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-qkg8t" podStartSLOduration=126.467724343 podStartE2EDuration="2m6.467724343s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.461279402 +0000 UTC m=+145.111547034" watchObservedRunningTime="2025-12-06 05:41:45.467724343 +0000 UTC m=+145.117991975" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.468341 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5b8p9" podStartSLOduration=126.46833441 podStartE2EDuration="2m6.46833441s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.377999664 +0000 UTC m=+145.028267296" watchObservedRunningTime="2025-12-06 05:41:45.46833441 +0000 UTC m=+145.118602042" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.522870 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.523236 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.023220617 +0000 UTC m=+145.673488249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.524016 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-fls7b" podStartSLOduration=126.524001539 podStartE2EDuration="2m6.524001539s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.515034537 +0000 UTC m=+145.165302189" watchObservedRunningTime="2025-12-06 05:41:45.524001539 +0000 UTC m=+145.174269181" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.556349 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-284vj" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.624513 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.625130 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.125083428 +0000 UTC m=+145.775351060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.625800 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.626578 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.1265592 +0000 UTC m=+145.776826842 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.647094 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-jbd4c" podStartSLOduration=126.647039537 podStartE2EDuration="2m6.647039537s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.643807436 +0000 UTC m=+145.294075078" watchObservedRunningTime="2025-12-06 05:41:45.647039537 +0000 UTC m=+145.297307189" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.685518 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" podStartSLOduration=126.685488991 podStartE2EDuration="2m6.685488991s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.68014055 +0000 UTC m=+145.330408202" watchObservedRunningTime="2025-12-06 05:41:45.685488991 +0000 UTC m=+145.335756623" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.709266 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8stt7" podStartSLOduration=126.70923987 podStartE2EDuration="2m6.70923987s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.708183121 +0000 UTC m=+145.358450753" watchObservedRunningTime="2025-12-06 05:41:45.70923987 +0000 UTC m=+145.359507502" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.727702 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.728220 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.228197185 +0000 UTC m=+145.878464817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.780678 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nz6jh" podStartSLOduration=126.780655394 podStartE2EDuration="2m6.780655394s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.780130539 +0000 UTC m=+145.430398191" watchObservedRunningTime="2025-12-06 05:41:45.780655394 +0000 UTC m=+145.430923016" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.781959 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m7xn8" podStartSLOduration=126.78195199 podStartE2EDuration="2m6.78195199s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.741251153 +0000 UTC m=+145.391518795" watchObservedRunningTime="2025-12-06 05:41:45.78195199 +0000 UTC m=+145.432219622" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.829460 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.829808 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.329794858 +0000 UTC m=+145.980062490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.876419 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9ckdj" podStartSLOduration=126.876404222 podStartE2EDuration="2m6.876404222s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.874812917 +0000 UTC m=+145.525080559" watchObservedRunningTime="2025-12-06 05:41:45.876404222 +0000 UTC m=+145.526671854" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.913186 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-k4xws" podStartSLOduration=126.913170729 podStartE2EDuration="2m6.913170729s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:45.911701218 +0000 UTC m=+145.561968860" watchObservedRunningTime="2025-12-06 05:41:45.913170729 +0000 UTC m=+145.563438361" Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.923449 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" event={"ID":"aa443eb3-221f-42d3-a946-f5bfee0d7148","Type":"ContainerStarted","Data":"c487e0867786b3481405a58a73696f17a8c0bf5cf67c0908b35788d2d3f0f1f7"} Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.930114 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:45 crc kubenswrapper[4957]: E1206 05:41:45.930733 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.430717653 +0000 UTC m=+146.080985285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.957323 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" event={"ID":"cd29b564-939e-4304-9f8e-11aefcb02a5e","Type":"ContainerStarted","Data":"8ef876fb797e2b46220ebc04f7d3b572b9e8ed5b6f62bb37ed8ac9396d5350b2"} Dec 06 05:41:45 crc kubenswrapper[4957]: I1206 05:41:45.969213 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.035776 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" podStartSLOduration=127.035743684 podStartE2EDuration="2m7.035743684s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:46.004489113 +0000 UTC m=+145.654756765" watchObservedRunningTime="2025-12-06 05:41:46.035743684 +0000 UTC m=+145.686011316" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.036800 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.062116 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.562100016 +0000 UTC m=+146.212367848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.139526 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.139911 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.639886219 +0000 UTC m=+146.290153851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.140202 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.140567 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.640559028 +0000 UTC m=+146.290826660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.200213 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.200278 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.214127 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.241495 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.242523 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.742508842 +0000 UTC m=+146.392776464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.284467 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:46 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:46 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:46 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.284526 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.343135 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.343791 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.843773576 +0000 UTC m=+146.494041208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.347489 4957 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftjwd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.347703 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ftjwd" podUID="c9105dc1-76e5-46dd-854a-c79f1c000c27" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.348033 4957 patch_prober.go:28] interesting pod/downloads-7954f5f757-ftjwd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.348066 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ftjwd" podUID="c9105dc1-76e5-46dd-854a-c79f1c000c27" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.365288 4957 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.445092 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.445463 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.945419641 +0000 UTC m=+146.595687273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.445657 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.445982 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:46.945968296 +0000 UTC m=+146.596235928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.479307 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-44rx8"] Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.480251 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.483982 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.510209 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-44rx8"] Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.546763 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.547157 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-catalog-content\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.547272 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-utilities\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.547321 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brs4f\" (UniqueName: \"kubernetes.io/projected/0dc111dd-bc98-4dfc-8130-85c33d2271b5-kube-api-access-brs4f\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.547540 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:47.047507579 +0000 UTC m=+146.697775221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.648250 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-utilities\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.649146 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.649091 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-utilities\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.649462 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:47.149450312 +0000 UTC m=+146.799717944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.649777 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brs4f\" (UniqueName: \"kubernetes.io/projected/0dc111dd-bc98-4dfc-8130-85c33d2271b5-kube-api-access-brs4f\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.650124 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-catalog-content\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.650421 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-catalog-content\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.671562 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kbk94"] Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.672756 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.675562 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brs4f\" (UniqueName: \"kubernetes.io/projected/0dc111dd-bc98-4dfc-8130-85c33d2271b5-kube-api-access-brs4f\") pod \"certified-operators-44rx8\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.677144 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.682393 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbk94"] Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.708258 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.708387 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.709682 4957 patch_prober.go:28] interesting pod/console-f9d7485db-8mb7l container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.37:8443/health\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.709758 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8mb7l" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerName="console" probeResult="failure" output="Get \"https://10.217.0.37:8443/health\": dial tcp 10.217.0.37:8443: connect: connection refused" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.740930 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.741046 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.751139 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.751351 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 05:41:47.251310662 +0000 UTC m=+146.901578294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.751630 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckm4d\" (UniqueName: \"kubernetes.io/projected/5de294a0-fcab-43a3-8001-c32c3965f3f6-kube-api-access-ckm4d\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.751822 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-utilities\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.752134 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-catalog-content\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.752219 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:46 crc kubenswrapper[4957]: E1206 05:41:46.752968 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 05:41:47.252959989 +0000 UTC m=+146.903227621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xlz55" (UID: "8c269452-c147-4adc-9280-7dd91cff2687") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.793695 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.821511 4957 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T05:41:46.365319224Z","Handler":null,"Name":""} Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.824898 4957 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.824956 4957 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.854164 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.854863 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.855170 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-catalog-content\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.855235 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.855294 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.855341 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckm4d\" (UniqueName: \"kubernetes.io/projected/5de294a0-fcab-43a3-8001-c32c3965f3f6-kube-api-access-ckm4d\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.855388 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-utilities\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.855437 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.855671 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-catalog-content\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.856411 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-utilities\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.858402 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.858772 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.864555 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.866263 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.885579 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckm4d\" (UniqueName: \"kubernetes.io/projected/5de294a0-fcab-43a3-8001-c32c3965f3f6-kube-api-access-ckm4d\") pod \"community-operators-kbk94\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.889225 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hj4gr"] Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.890170 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.915261 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hj4gr"] Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.956484 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-catalog-content\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.956545 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-utilities\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.956574 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfkfz\" (UniqueName: \"kubernetes.io/projected/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-kube-api-access-gfkfz\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:46 crc kubenswrapper[4957]: I1206 05:41:46.956621 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.007015 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" event={"ID":"aa443eb3-221f-42d3-a946-f5bfee0d7148","Type":"ContainerStarted","Data":"a79e1802fe240c2fe1cf4a86275ca2b21e70265be071cc25af07ec7ab46d7872"} Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.007075 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" event={"ID":"aa443eb3-221f-42d3-a946-f5bfee0d7148","Type":"ContainerStarted","Data":"5004c066f1e8a8e321c082aba9eb38a3af6d8dd9487ca346da4dc91c35c90dc1"} Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.025618 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.025981 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7qkmz" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.029946 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9zmkv" podStartSLOduration=13.029929316 podStartE2EDuration="13.029929316s" podCreationTimestamp="2025-12-06 05:41:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:47.029538155 +0000 UTC m=+146.679805817" watchObservedRunningTime="2025-12-06 05:41:47.029929316 +0000 UTC m=+146.680196948" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.058764 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfkfz\" (UniqueName: \"kubernetes.io/projected/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-kube-api-access-gfkfz\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.059251 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-catalog-content\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.059359 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-utilities\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.067132 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-catalog-content\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.071202 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-utilities\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.088884 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-44rx8"] Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.093546 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfkfz\" (UniqueName: \"kubernetes.io/projected/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-kube-api-access-gfkfz\") pod \"certified-operators-hj4gr\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.102173 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.102783 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-glf2z"] Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.104513 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.107492 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glf2z"] Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.109704 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.171152 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtqzx\" (UniqueName: \"kubernetes.io/projected/88ca26d1-ec5d-44ee-9bae-db531711bd72-kube-api-access-xtqzx\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.171214 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-catalog-content\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.171266 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-utilities\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.209389 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.229883 4957 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.229924 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.232630 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.274013 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtqzx\" (UniqueName: \"kubernetes.io/projected/88ca26d1-ec5d-44ee-9bae-db531711bd72-kube-api-access-xtqzx\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.274112 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-catalog-content\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.274181 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-utilities\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.274780 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-utilities\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.275093 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-catalog-content\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.282929 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.289826 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:47 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.289929 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.304160 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtqzx\" (UniqueName: \"kubernetes.io/projected/88ca26d1-ec5d-44ee-9bae-db531711bd72-kube-api-access-xtqzx\") pod \"community-operators-glf2z\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.323204 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xlz55\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.409214 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbk94"] Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.420368 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 05:41:47 crc kubenswrapper[4957]: W1206 05:41:47.430265 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5de294a0_fcab_43a3_8001_c32c3965f3f6.slice/crio-f8f8db63e6e74dd360753249b7d6db258bbfe99ea610194b80d31db13da8b292 WatchSource:0}: Error finding container f8f8db63e6e74dd360753249b7d6db258bbfe99ea610194b80d31db13da8b292: Status 404 returned error can't find the container with id f8f8db63e6e74dd360753249b7d6db258bbfe99ea610194b80d31db13da8b292 Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.447430 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.531027 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.787278 4957 patch_prober.go:28] interesting pod/apiserver-76f77b778f-hj5wv container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 05:41:47 crc kubenswrapper[4957]: [+]log ok Dec 06 05:41:47 crc kubenswrapper[4957]: [+]etcd ok Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/start-apiserver-admission-initializer failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/generic-apiserver-start-informers failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/max-in-flight-filter failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/storage-object-count-tracker-hook failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/image.openshift.io-apiserver-caches failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [+]poststarthook/project.openshift.io-projectcache ok Dec 06 05:41:47 crc kubenswrapper[4957]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/openshift.io-restmapperupdater failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: [-]poststarthook/quota.openshift.io-clusterquotamapping failed: reason withheld Dec 06 05:41:47 crc kubenswrapper[4957]: livez check failed Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.787799 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" podUID="cd29b564-939e-4304-9f8e-11aefcb02a5e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.907995 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hj4gr"] Dec 06 05:41:47 crc kubenswrapper[4957]: W1206 05:41:47.961024 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87e7bb85_760f_4c78_99b8_6a8b38e8f70f.slice/crio-cf55a42b23539cfad5d1b63acd31d43ced8702c31830c6cf37a8f444e3701e19 WatchSource:0}: Error finding container cf55a42b23539cfad5d1b63acd31d43ced8702c31830c6cf37a8f444e3701e19: Status 404 returned error can't find the container with id cf55a42b23539cfad5d1b63acd31d43ced8702c31830c6cf37a8f444e3701e19 Dec 06 05:41:47 crc kubenswrapper[4957]: I1206 05:41:47.965297 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glf2z"] Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.014948 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"53b2dee6d0b16f695d28fb25023916ac3e56bf3d3e13fa366d33f38e10c5f37e"} Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.018246 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a8882e473028bfe1a3a794adcd6904faaf84ee4039d5e3ddcec4f447c76d05ef"} Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.019421 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"546e12efcbf7ea64b258d8e24b76e2e69c786041779b1a90325f1a7778eb60ff"} Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.020484 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbk94" event={"ID":"5de294a0-fcab-43a3-8001-c32c3965f3f6","Type":"ContainerStarted","Data":"f8f8db63e6e74dd360753249b7d6db258bbfe99ea610194b80d31db13da8b292"} Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.025464 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hj4gr" event={"ID":"87e7bb85-760f-4c78-99b8-6a8b38e8f70f","Type":"ContainerStarted","Data":"cf55a42b23539cfad5d1b63acd31d43ced8702c31830c6cf37a8f444e3701e19"} Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.027341 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44rx8" event={"ID":"0dc111dd-bc98-4dfc-8130-85c33d2271b5","Type":"ContainerStarted","Data":"9ca1b4f104365c6c037684aeb9acad1fe793b27a68e4bd0b106834530fbd601e"} Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.029719 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glf2z" event={"ID":"88ca26d1-ec5d-44ee-9bae-db531711bd72","Type":"ContainerStarted","Data":"de34137f3316b417536fea72799f77996cc2a2eec4b732f12f7562e4de16086a"} Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.197059 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xlz55"] Dec 06 05:41:48 crc kubenswrapper[4957]: W1206 05:41:48.206051 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c269452_c147_4adc_9280_7dd91cff2687.slice/crio-171d583a4ba9f746d1290bffdba8fb0ca8bc3ecb0dba69eef857fd88b366cfea WatchSource:0}: Error finding container 171d583a4ba9f746d1290bffdba8fb0ca8bc3ecb0dba69eef857fd88b366cfea: Status 404 returned error can't find the container with id 171d583a4ba9f746d1290bffdba8fb0ca8bc3ecb0dba69eef857fd88b366cfea Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.286537 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:48 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:48 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:48 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.286602 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.669647 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.860401 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2f5bt"] Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.861398 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.863653 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.872410 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2f5bt"] Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.897795 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-utilities\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.897847 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-catalog-content\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.897879 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n2sg\" (UniqueName: \"kubernetes.io/projected/37639dbf-8f08-428e-93e2-456e2946333f-kube-api-access-5n2sg\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.998654 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-utilities\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.999336 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-catalog-content\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.999270 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-utilities\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.999429 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n2sg\" (UniqueName: \"kubernetes.io/projected/37639dbf-8f08-428e-93e2-456e2946333f-kube-api-access-5n2sg\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:48 crc kubenswrapper[4957]: I1206 05:41:48.999879 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-catalog-content\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.018694 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n2sg\" (UniqueName: \"kubernetes.io/projected/37639dbf-8f08-428e-93e2-456e2946333f-kube-api-access-5n2sg\") pod \"redhat-marketplace-2f5bt\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.036103 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" event={"ID":"8c269452-c147-4adc-9280-7dd91cff2687","Type":"ContainerStarted","Data":"386075d3ff234a92cadf00ef8864691d2f2c9ed3aa7012520e0111ced9c791aa"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.036225 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" event={"ID":"8c269452-c147-4adc-9280-7dd91cff2687","Type":"ContainerStarted","Data":"171d583a4ba9f746d1290bffdba8fb0ca8bc3ecb0dba69eef857fd88b366cfea"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.036249 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.037861 4957 generic.go:334] "Generic (PLEG): container finished" podID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerID="6a64e7fd6b23bcfd66f652ca8510bb86c2227cd72c2960ae989551eb68d324b0" exitCode=0 Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.037890 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44rx8" event={"ID":"0dc111dd-bc98-4dfc-8130-85c33d2271b5","Type":"ContainerDied","Data":"6a64e7fd6b23bcfd66f652ca8510bb86c2227cd72c2960ae989551eb68d324b0"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.039464 4957 generic.go:334] "Generic (PLEG): container finished" podID="f9af915f-4e5b-440d-a3b3-534d16b1f686" containerID="cef6bc9892b207b2f3b961a1b79ffebf77071515014aeeb07dbe38b4fc6a239c" exitCode=0 Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.039502 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" event={"ID":"f9af915f-4e5b-440d-a3b3-534d16b1f686","Type":"ContainerDied","Data":"cef6bc9892b207b2f3b961a1b79ffebf77071515014aeeb07dbe38b4fc6a239c"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.039897 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.042565 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7142d58e83d81042175dfd013c552c1810bba1ee6964504cb7adfc9014d247ce"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.042698 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.043772 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d3be471540ffdd575c72503d8285e7db0d56f7ae5f3e94c520ecd5cd6b7a98ae"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.053265 4957 generic.go:334] "Generic (PLEG): container finished" podID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerID="57c33eb9f88adc9309321905995d7c3532105bf50b257194b8ed8005a53d4890" exitCode=0 Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.053612 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbk94" event={"ID":"5de294a0-fcab-43a3-8001-c32c3965f3f6","Type":"ContainerDied","Data":"57c33eb9f88adc9309321905995d7c3532105bf50b257194b8ed8005a53d4890"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.056442 4957 generic.go:334] "Generic (PLEG): container finished" podID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerID="c6f95c6ceebca8842826a5a7327239d6fdf3993ce68f9b84321687180efd4b6b" exitCode=0 Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.056512 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hj4gr" event={"ID":"87e7bb85-760f-4c78-99b8-6a8b38e8f70f","Type":"ContainerDied","Data":"c6f95c6ceebca8842826a5a7327239d6fdf3993ce68f9b84321687180efd4b6b"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.060494 4957 generic.go:334] "Generic (PLEG): container finished" podID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerID="f7a11a3d0bcbbea2a8b00ebd1141065f433835fb4f0c621d4e06a1819922961d" exitCode=0 Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.060531 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glf2z" event={"ID":"88ca26d1-ec5d-44ee-9bae-db531711bd72","Type":"ContainerDied","Data":"f7a11a3d0bcbbea2a8b00ebd1141065f433835fb4f0c621d4e06a1819922961d"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.061970 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"29a3367e33f99aa78691275acc559139e66bb37b49a015b00c76cf548192a9cb"} Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.064059 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" podStartSLOduration=130.06403481 podStartE2EDuration="2m10.06403481s" podCreationTimestamp="2025-12-06 05:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:41:49.063413063 +0000 UTC m=+148.713680705" watchObservedRunningTime="2025-12-06 05:41:49.06403481 +0000 UTC m=+148.714302442" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.179976 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.266421 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g5chm"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.267726 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.280447 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5chm"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.296776 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:49 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:49 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:49 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.296825 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.408631 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2f5bt"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.413043 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-567zz\" (UniqueName: \"kubernetes.io/projected/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-kube-api-access-567zz\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.413097 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-catalog-content\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.413172 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-utilities\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: W1206 05:41:49.427020 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37639dbf_8f08_428e_93e2_456e2946333f.slice/crio-462b207bd97a6c6d30b332b360e18ac92dbc7fda833b6338f5eec0ff80bd41f3 WatchSource:0}: Error finding container 462b207bd97a6c6d30b332b360e18ac92dbc7fda833b6338f5eec0ff80bd41f3: Status 404 returned error can't find the container with id 462b207bd97a6c6d30b332b360e18ac92dbc7fda833b6338f5eec0ff80bd41f3 Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.514825 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-567zz\" (UniqueName: \"kubernetes.io/projected/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-kube-api-access-567zz\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.514878 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-catalog-content\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.514950 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-utilities\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.515378 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-utilities\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.515570 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-catalog-content\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.535589 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-567zz\" (UniqueName: \"kubernetes.io/projected/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-kube-api-access-567zz\") pod \"redhat-marketplace-g5chm\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.601257 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.658514 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mzt5s"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.659455 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.662309 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.667554 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mzt5s"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.809487 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5chm"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.818458 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-utilities\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.818534 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrggc\" (UniqueName: \"kubernetes.io/projected/f6be4201-af3d-4f37-a819-7cf136f1e906-kube-api-access-mrggc\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.818623 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-catalog-content\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: W1206 05:41:49.820319 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2e65b89_5a7e_4845_9bd7_6ae4c3e936f4.slice/crio-6ad7804cbab340b5c2d676930e44abf947179031754416ae161f3c70924d2e0b WatchSource:0}: Error finding container 6ad7804cbab340b5c2d676930e44abf947179031754416ae161f3c70924d2e0b: Status 404 returned error can't find the container with id 6ad7804cbab340b5c2d676930e44abf947179031754416ae161f3c70924d2e0b Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.863721 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hz4nt"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.865526 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.874081 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hz4nt"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.919332 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-catalog-content\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.919406 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-utilities\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.919423 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrggc\" (UniqueName: \"kubernetes.io/projected/f6be4201-af3d-4f37-a819-7cf136f1e906-kube-api-access-mrggc\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.920106 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-catalog-content\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.920280 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-utilities\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.938192 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrggc\" (UniqueName: \"kubernetes.io/projected/f6be4201-af3d-4f37-a819-7cf136f1e906-kube-api-access-mrggc\") pod \"redhat-operators-mzt5s\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.964557 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.975176 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.977732 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.981623 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.983209 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 05:41:49 crc kubenswrapper[4957]: I1206 05:41:49.984033 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.020542 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-catalog-content\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.020750 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjgb6\" (UniqueName: \"kubernetes.io/projected/b5b5df82-a892-4971-84e3-3c3690960f71-kube-api-access-pjgb6\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.020891 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-utilities\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.076744 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5chm" event={"ID":"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4","Type":"ContainerStarted","Data":"6ad7804cbab340b5c2d676930e44abf947179031754416ae161f3c70924d2e0b"} Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.078368 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2f5bt" event={"ID":"37639dbf-8f08-428e-93e2-456e2946333f","Type":"ContainerStarted","Data":"462b207bd97a6c6d30b332b360e18ac92dbc7fda833b6338f5eec0ff80bd41f3"} Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.122067 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-utilities\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.122186 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-catalog-content\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.122254 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.122416 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjgb6\" (UniqueName: \"kubernetes.io/projected/b5b5df82-a892-4971-84e3-3c3690960f71-kube-api-access-pjgb6\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.122453 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.122640 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-utilities\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.124111 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-catalog-content\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.143587 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjgb6\" (UniqueName: \"kubernetes.io/projected/b5b5df82-a892-4971-84e3-3c3690960f71-kube-api-access-pjgb6\") pod \"redhat-operators-hz4nt\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.170426 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mzt5s"] Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.196649 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.223470 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.223570 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.225662 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.240172 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.274186 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.294907 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:50 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:50 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:50 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.294972 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.315178 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.417975 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hz4nt"] Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.425498 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9af915f-4e5b-440d-a3b3-534d16b1f686-secret-volume\") pod \"f9af915f-4e5b-440d-a3b3-534d16b1f686\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.425787 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9af915f-4e5b-440d-a3b3-534d16b1f686-config-volume\") pod \"f9af915f-4e5b-440d-a3b3-534d16b1f686\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.425817 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ps2f\" (UniqueName: \"kubernetes.io/projected/f9af915f-4e5b-440d-a3b3-534d16b1f686-kube-api-access-4ps2f\") pod \"f9af915f-4e5b-440d-a3b3-534d16b1f686\" (UID: \"f9af915f-4e5b-440d-a3b3-534d16b1f686\") " Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.426555 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9af915f-4e5b-440d-a3b3-534d16b1f686-config-volume" (OuterVolumeSpecName: "config-volume") pod "f9af915f-4e5b-440d-a3b3-534d16b1f686" (UID: "f9af915f-4e5b-440d-a3b3-534d16b1f686"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.426880 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9af915f-4e5b-440d-a3b3-534d16b1f686-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.430983 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9af915f-4e5b-440d-a3b3-534d16b1f686-kube-api-access-4ps2f" (OuterVolumeSpecName: "kube-api-access-4ps2f") pod "f9af915f-4e5b-440d-a3b3-534d16b1f686" (UID: "f9af915f-4e5b-440d-a3b3-534d16b1f686"). InnerVolumeSpecName "kube-api-access-4ps2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.431339 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9af915f-4e5b-440d-a3b3-534d16b1f686-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f9af915f-4e5b-440d-a3b3-534d16b1f686" (UID: "f9af915f-4e5b-440d-a3b3-534d16b1f686"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.528550 4957 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9af915f-4e5b-440d-a3b3-534d16b1f686-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.528593 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ps2f\" (UniqueName: \"kubernetes.io/projected/f9af915f-4e5b-440d-a3b3-534d16b1f686-kube-api-access-4ps2f\") on node \"crc\" DevicePath \"\"" Dec 06 05:41:50 crc kubenswrapper[4957]: I1206 05:41:50.533597 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 05:41:50 crc kubenswrapper[4957]: W1206 05:41:50.544819 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode86e9e17_e4cc_4c66_97a8_365f78e899ff.slice/crio-6baf603c3b27f9ff4ee927f9dce0928291de589cd6d29ecfa0218ff4381a02c2 WatchSource:0}: Error finding container 6baf603c3b27f9ff4ee927f9dce0928291de589cd6d29ecfa0218ff4381a02c2: Status 404 returned error can't find the container with id 6baf603c3b27f9ff4ee927f9dce0928291de589cd6d29ecfa0218ff4381a02c2 Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.090875 4957 generic.go:334] "Generic (PLEG): container finished" podID="37639dbf-8f08-428e-93e2-456e2946333f" containerID="3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211" exitCode=0 Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.091297 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2f5bt" event={"ID":"37639dbf-8f08-428e-93e2-456e2946333f","Type":"ContainerDied","Data":"3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.094867 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e86e9e17-e4cc-4c66-97a8-365f78e899ff","Type":"ContainerStarted","Data":"6baf603c3b27f9ff4ee927f9dce0928291de589cd6d29ecfa0218ff4381a02c2"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.116557 4957 generic.go:334] "Generic (PLEG): container finished" podID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerID="7addc9291f4aa344f94265482364f8ed0bf3c79f5631d14c262262e61ec43090" exitCode=0 Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.116672 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5chm" event={"ID":"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4","Type":"ContainerDied","Data":"7addc9291f4aa344f94265482364f8ed0bf3c79f5631d14c262262e61ec43090"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.119228 4957 generic.go:334] "Generic (PLEG): container finished" podID="b5b5df82-a892-4971-84e3-3c3690960f71" containerID="b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab" exitCode=0 Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.119292 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hz4nt" event={"ID":"b5b5df82-a892-4971-84e3-3c3690960f71","Type":"ContainerDied","Data":"b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.119319 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hz4nt" event={"ID":"b5b5df82-a892-4971-84e3-3c3690960f71","Type":"ContainerStarted","Data":"5cbff16ccfe7539ba8ca31e6ad2be89bd640862e8460ade6beb112919c7b2e70"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.129942 4957 generic.go:334] "Generic (PLEG): container finished" podID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerID="dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4" exitCode=0 Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.130014 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzt5s" event={"ID":"f6be4201-af3d-4f37-a819-7cf136f1e906","Type":"ContainerDied","Data":"dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.130046 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzt5s" event={"ID":"f6be4201-af3d-4f37-a819-7cf136f1e906","Type":"ContainerStarted","Data":"4d12c532ee11e1fea2995921d7685e330b63c5c2a1a7c5f5efd2561f6b2fc714"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.141378 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" event={"ID":"f9af915f-4e5b-440d-a3b3-534d16b1f686","Type":"ContainerDied","Data":"de574bc3179f1507d2559b80e1cf75f068c151c57445cc4a6aa42f7e84051298"} Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.141420 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.141425 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de574bc3179f1507d2559b80e1cf75f068c151c57445cc4a6aa42f7e84051298" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.285400 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:51 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:51 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:51 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.285485 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.567351 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 05:41:51 crc kubenswrapper[4957]: E1206 05:41:51.567923 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9af915f-4e5b-440d-a3b3-534d16b1f686" containerName="collect-profiles" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.568045 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9af915f-4e5b-440d-a3b3-534d16b1f686" containerName="collect-profiles" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.568231 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9af915f-4e5b-440d-a3b3-534d16b1f686" containerName="collect-profiles" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.568757 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.571690 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.572073 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.573704 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.741417 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78137157-1171-4b4a-bf29-dd1162066a5e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.741529 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78137157-1171-4b4a-bf29-dd1162066a5e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.744490 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.748955 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hj5wv" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.844470 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78137157-1171-4b4a-bf29-dd1162066a5e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.844657 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78137157-1171-4b4a-bf29-dd1162066a5e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.845977 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78137157-1171-4b4a-bf29-dd1162066a5e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.871374 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78137157-1171-4b4a-bf29-dd1162066a5e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:51 crc kubenswrapper[4957]: I1206 05:41:51.889225 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:52 crc kubenswrapper[4957]: I1206 05:41:52.114255 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 05:41:52 crc kubenswrapper[4957]: I1206 05:41:52.167412 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"78137157-1171-4b4a-bf29-dd1162066a5e","Type":"ContainerStarted","Data":"0b5b3cbe17d4afa9f8768ac596128ec6ebd69471fb8aef63abfaaa4cce7da8c3"} Dec 06 05:41:52 crc kubenswrapper[4957]: I1206 05:41:52.284731 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:52 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:52 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:52 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:52 crc kubenswrapper[4957]: I1206 05:41:52.285037 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:53 crc kubenswrapper[4957]: I1206 05:41:53.178440 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e86e9e17-e4cc-4c66-97a8-365f78e899ff","Type":"ContainerStarted","Data":"de3ea0824f1f0b408120efde8d6290627252ed6ef18ff2e9cbebc7758b6bc846"} Dec 06 05:41:53 crc kubenswrapper[4957]: I1206 05:41:53.285513 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:53 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:53 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:53 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:53 crc kubenswrapper[4957]: I1206 05:41:53.285602 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:54 crc kubenswrapper[4957]: I1206 05:41:54.184227 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"78137157-1171-4b4a-bf29-dd1162066a5e","Type":"ContainerStarted","Data":"801f19077f72def41e66444a636bc4a214199eae2ec7b0447c099b1e687e7c9f"} Dec 06 05:41:54 crc kubenswrapper[4957]: I1206 05:41:54.284596 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:54 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:54 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:54 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:54 crc kubenswrapper[4957]: I1206 05:41:54.284674 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:54 crc kubenswrapper[4957]: I1206 05:41:54.721771 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-tbzz4" Dec 06 05:41:55 crc kubenswrapper[4957]: I1206 05:41:55.191545 4957 generic.go:334] "Generic (PLEG): container finished" podID="78137157-1171-4b4a-bf29-dd1162066a5e" containerID="801f19077f72def41e66444a636bc4a214199eae2ec7b0447c099b1e687e7c9f" exitCode=0 Dec 06 05:41:55 crc kubenswrapper[4957]: I1206 05:41:55.191614 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"78137157-1171-4b4a-bf29-dd1162066a5e","Type":"ContainerDied","Data":"801f19077f72def41e66444a636bc4a214199eae2ec7b0447c099b1e687e7c9f"} Dec 06 05:41:55 crc kubenswrapper[4957]: I1206 05:41:55.202548 4957 generic.go:334] "Generic (PLEG): container finished" podID="e86e9e17-e4cc-4c66-97a8-365f78e899ff" containerID="de3ea0824f1f0b408120efde8d6290627252ed6ef18ff2e9cbebc7758b6bc846" exitCode=0 Dec 06 05:41:55 crc kubenswrapper[4957]: I1206 05:41:55.202618 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e86e9e17-e4cc-4c66-97a8-365f78e899ff","Type":"ContainerDied","Data":"de3ea0824f1f0b408120efde8d6290627252ed6ef18ff2e9cbebc7758b6bc846"} Dec 06 05:41:55 crc kubenswrapper[4957]: I1206 05:41:55.285055 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:55 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:55 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:55 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:55 crc kubenswrapper[4957]: I1206 05:41:55.285114 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.283786 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:56 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:56 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:56 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.284152 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.355462 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ftjwd" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.538084 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.540505 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.708602 4957 patch_prober.go:28] interesting pod/console-f9d7485db-8mb7l container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.37:8443/health\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.708698 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8mb7l" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerName="console" probeResult="failure" output="Get \"https://10.217.0.37:8443/health\": dial tcp 10.217.0.37:8443: connect: connection refused" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716061 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78137157-1171-4b4a-bf29-dd1162066a5e-kube-api-access\") pod \"78137157-1171-4b4a-bf29-dd1162066a5e\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716118 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78137157-1171-4b4a-bf29-dd1162066a5e-kubelet-dir\") pod \"78137157-1171-4b4a-bf29-dd1162066a5e\" (UID: \"78137157-1171-4b4a-bf29-dd1162066a5e\") " Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716162 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kube-api-access\") pod \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716213 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kubelet-dir\") pod \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\" (UID: \"e86e9e17-e4cc-4c66-97a8-365f78e899ff\") " Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716255 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78137157-1171-4b4a-bf29-dd1162066a5e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "78137157-1171-4b4a-bf29-dd1162066a5e" (UID: "78137157-1171-4b4a-bf29-dd1162066a5e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716385 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e86e9e17-e4cc-4c66-97a8-365f78e899ff" (UID: "e86e9e17-e4cc-4c66-97a8-365f78e899ff"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716698 4957 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/78137157-1171-4b4a-bf29-dd1162066a5e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.716737 4957 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.722560 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e86e9e17-e4cc-4c66-97a8-365f78e899ff" (UID: "e86e9e17-e4cc-4c66-97a8-365f78e899ff"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.723098 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78137157-1171-4b4a-bf29-dd1162066a5e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "78137157-1171-4b4a-bf29-dd1162066a5e" (UID: "78137157-1171-4b4a-bf29-dd1162066a5e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.818204 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e86e9e17-e4cc-4c66-97a8-365f78e899ff-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:41:56 crc kubenswrapper[4957]: I1206 05:41:56.818587 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78137157-1171-4b4a-bf29-dd1162066a5e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.225880 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.225859 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"78137157-1171-4b4a-bf29-dd1162066a5e","Type":"ContainerDied","Data":"0b5b3cbe17d4afa9f8768ac596128ec6ebd69471fb8aef63abfaaa4cce7da8c3"} Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.226025 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b5b3cbe17d4afa9f8768ac596128ec6ebd69471fb8aef63abfaaa4cce7da8c3" Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.227391 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e86e9e17-e4cc-4c66-97a8-365f78e899ff","Type":"ContainerDied","Data":"6baf603c3b27f9ff4ee927f9dce0928291de589cd6d29ecfa0218ff4381a02c2"} Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.227433 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6baf603c3b27f9ff4ee927f9dce0928291de589cd6d29ecfa0218ff4381a02c2" Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.227490 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.284236 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:57 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:57 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:57 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:57 crc kubenswrapper[4957]: I1206 05:41:57.284301 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:58 crc kubenswrapper[4957]: I1206 05:41:58.283384 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:58 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:58 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:58 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:58 crc kubenswrapper[4957]: I1206 05:41:58.283440 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:41:58 crc kubenswrapper[4957]: I1206 05:41:58.547213 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:41:59 crc kubenswrapper[4957]: I1206 05:41:59.283355 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:41:59 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:41:59 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:41:59 crc kubenswrapper[4957]: healthz check failed Dec 06 05:41:59 crc kubenswrapper[4957]: I1206 05:41:59.283617 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:42:00 crc kubenswrapper[4957]: I1206 05:42:00.285158 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:42:00 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:42:00 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:42:00 crc kubenswrapper[4957]: healthz check failed Dec 06 05:42:00 crc kubenswrapper[4957]: I1206 05:42:00.285492 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:42:01 crc kubenswrapper[4957]: I1206 05:42:01.283500 4957 patch_prober.go:28] interesting pod/router-default-5444994796-9kjdh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 05:42:01 crc kubenswrapper[4957]: [-]has-synced failed: reason withheld Dec 06 05:42:01 crc kubenswrapper[4957]: [+]process-running ok Dec 06 05:42:01 crc kubenswrapper[4957]: healthz check failed Dec 06 05:42:01 crc kubenswrapper[4957]: I1206 05:42:01.283564 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9kjdh" podUID="f9ef9416-d92a-401b-8f80-81ab25b24d31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:42:02 crc kubenswrapper[4957]: I1206 05:42:02.105495 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:42:02 crc kubenswrapper[4957]: I1206 05:42:02.111264 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcf7c4c9-894b-450c-83f6-441d3639f75a-metrics-certs\") pod \"network-metrics-daemon-slqnh\" (UID: \"fcf7c4c9-894b-450c-83f6-441d3639f75a\") " pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:42:02 crc kubenswrapper[4957]: I1206 05:42:02.284417 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:42:02 crc kubenswrapper[4957]: I1206 05:42:02.287025 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9kjdh" Dec 06 05:42:02 crc kubenswrapper[4957]: I1206 05:42:02.389654 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-slqnh" Dec 06 05:42:06 crc kubenswrapper[4957]: I1206 05:42:06.730483 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:42:06 crc kubenswrapper[4957]: I1206 05:42:06.734033 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:42:07 crc kubenswrapper[4957]: I1206 05:42:07.537811 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:42:10 crc kubenswrapper[4957]: I1206 05:42:10.180727 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:42:10 crc kubenswrapper[4957]: I1206 05:42:10.180810 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:42:17 crc kubenswrapper[4957]: I1206 05:42:17.085744 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9xm5h" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.567485 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 05:42:25 crc kubenswrapper[4957]: E1206 05:42:25.568135 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78137157-1171-4b4a-bf29-dd1162066a5e" containerName="pruner" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.568157 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="78137157-1171-4b4a-bf29-dd1162066a5e" containerName="pruner" Dec 06 05:42:25 crc kubenswrapper[4957]: E1206 05:42:25.568190 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86e9e17-e4cc-4c66-97a8-365f78e899ff" containerName="pruner" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.568201 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86e9e17-e4cc-4c66-97a8-365f78e899ff" containerName="pruner" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.568358 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86e9e17-e4cc-4c66-97a8-365f78e899ff" containerName="pruner" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.568385 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="78137157-1171-4b4a-bf29-dd1162066a5e" containerName="pruner" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.568944 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.570735 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.574031 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.574284 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.676167 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb54e52a-e54e-4d2a-9431-872769baf100-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.676500 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb54e52a-e54e-4d2a-9431-872769baf100-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.777819 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb54e52a-e54e-4d2a-9431-872769baf100-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.778018 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb54e52a-e54e-4d2a-9431-872769baf100-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.777925 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb54e52a-e54e-4d2a-9431-872769baf100-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.804014 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb54e52a-e54e-4d2a-9431-872769baf100-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:25 crc kubenswrapper[4957]: I1206 05:42:25.893810 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:28 crc kubenswrapper[4957]: I1206 05:42:28.162058 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 05:42:30 crc kubenswrapper[4957]: I1206 05:42:30.969108 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 05:42:30 crc kubenswrapper[4957]: I1206 05:42:30.970937 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.015109 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.172394 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-var-lock\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.172488 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.172602 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kube-api-access\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.273561 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-var-lock\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.273702 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.273722 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-var-lock\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.273795 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.273809 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kube-api-access\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.307816 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kube-api-access\") pod \"installer-9-crc\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:31 crc kubenswrapper[4957]: I1206 05:42:31.315404 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:42:40 crc kubenswrapper[4957]: I1206 05:42:40.180729 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:42:40 crc kubenswrapper[4957]: I1206 05:42:40.181557 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:42:40 crc kubenswrapper[4957]: I1206 05:42:40.181631 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:42:40 crc kubenswrapper[4957]: I1206 05:42:40.182470 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:42:40 crc kubenswrapper[4957]: I1206 05:42:40.182688 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3" gracePeriod=600 Dec 06 05:42:40 crc kubenswrapper[4957]: E1206 05:42:40.608181 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 05:42:40 crc kubenswrapper[4957]: E1206 05:42:40.608749 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gfkfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hj4gr_openshift-marketplace(87e7bb85-760f-4c78-99b8-6a8b38e8f70f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:40 crc kubenswrapper[4957]: E1206 05:42:40.610091 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hj4gr" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" Dec 06 05:42:41 crc kubenswrapper[4957]: E1206 05:42:41.547193 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 05:42:41 crc kubenswrapper[4957]: E1206 05:42:41.547451 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brs4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-44rx8_openshift-marketplace(0dc111dd-bc98-4dfc-8130-85c33d2271b5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:41 crc kubenswrapper[4957]: E1206 05:42:41.548620 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-44rx8" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" Dec 06 05:42:41 crc kubenswrapper[4957]: I1206 05:42:41.552545 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3" exitCode=0 Dec 06 05:42:41 crc kubenswrapper[4957]: I1206 05:42:41.553674 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3"} Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.125612 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-44rx8" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.125639 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hj4gr" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.202476 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.202627 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ckm4d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kbk94_openshift-marketplace(5de294a0-fcab-43a3-8001-c32c3965f3f6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.203965 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kbk94" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.277820 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.278115 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xtqzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-glf2z_openshift-marketplace(88ca26d1-ec5d-44ee-9bae-db531711bd72): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:43 crc kubenswrapper[4957]: E1206 05:42:43.284936 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-glf2z" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.383314 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kbk94" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.383383 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-glf2z" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.458031 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.458660 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n2sg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2f5bt_openshift-marketplace(37639dbf-8f08-428e-93e2-456e2946333f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.459879 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2f5bt" podUID="37639dbf-8f08-428e-93e2-456e2946333f" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.471662 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.472404 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-567zz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-g5chm_openshift-marketplace(e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:44 crc kubenswrapper[4957]: E1206 05:42:44.474453 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-g5chm" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" Dec 06 05:42:44 crc kubenswrapper[4957]: I1206 05:42:44.827229 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.821700 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2f5bt" podUID="37639dbf-8f08-428e-93e2-456e2946333f" Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.822486 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-g5chm" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.854698 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.855252 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pjgb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hz4nt_openshift-marketplace(b5b5df82-a892-4971-84e3-3c3690960f71): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.856917 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hz4nt" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.860376 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.860487 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mrggc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mzt5s_openshift-marketplace(f6be4201-af3d-4f37-a819-7cf136f1e906): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:42:47 crc kubenswrapper[4957]: E1206 05:42:47.862140 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mzt5s" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.340815 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.345716 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-slqnh"] Dec 06 05:42:48 crc kubenswrapper[4957]: W1206 05:42:48.362417 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcf7c4c9_894b_450c_83f6_441d3639f75a.slice/crio-f0bb9d445d547665d585d402e6f34d0b03fc26e590cb2edfc8f26b6cadec1668 WatchSource:0}: Error finding container f0bb9d445d547665d585d402e6f34d0b03fc26e590cb2edfc8f26b6cadec1668: Status 404 returned error can't find the container with id f0bb9d445d547665d585d402e6f34d0b03fc26e590cb2edfc8f26b6cadec1668 Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.590940 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e62592-5ba0-42ee-b9a0-41bfa6173821","Type":"ContainerStarted","Data":"09b4be6e7ae9d04f91a618f478fccc6074474cc2d7a60b155562bcae6d27b1d7"} Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.592479 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb54e52a-e54e-4d2a-9431-872769baf100","Type":"ContainerStarted","Data":"1c8f296e927917fea00c94076606c536bf420c650c1b55800aa4bd6824209413"} Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.592502 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb54e52a-e54e-4d2a-9431-872769baf100","Type":"ContainerStarted","Data":"dbf8ed06a0a149a8cee5fcc518997aa73d2ee172ad6308977286bc3748dee734"} Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.594477 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"7ffb51b23be1ebfd26c70f252394c0ca06f92ba61d22e53e094223c893ca1069"} Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.596170 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-slqnh" event={"ID":"fcf7c4c9-894b-450c-83f6-441d3639f75a","Type":"ContainerStarted","Data":"154c4d9da960dcf129fef513533c7d29b85df55d252f91597a822575de2f5865"} Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.596237 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-slqnh" event={"ID":"fcf7c4c9-894b-450c-83f6-441d3639f75a","Type":"ContainerStarted","Data":"f0bb9d445d547665d585d402e6f34d0b03fc26e590cb2edfc8f26b6cadec1668"} Dec 06 05:42:48 crc kubenswrapper[4957]: E1206 05:42:48.597390 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mzt5s" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" Dec 06 05:42:48 crc kubenswrapper[4957]: E1206 05:42:48.597574 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-hz4nt" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" Dec 06 05:42:48 crc kubenswrapper[4957]: I1206 05:42:48.614135 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=23.614115499 podStartE2EDuration="23.614115499s" podCreationTimestamp="2025-12-06 05:42:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:42:48.60603146 +0000 UTC m=+208.256299102" watchObservedRunningTime="2025-12-06 05:42:48.614115499 +0000 UTC m=+208.264383131" Dec 06 05:42:49 crc kubenswrapper[4957]: I1206 05:42:49.602981 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-slqnh" event={"ID":"fcf7c4c9-894b-450c-83f6-441d3639f75a","Type":"ContainerStarted","Data":"4662a1a17b5ce1b3b876557ca3161a90c9e7f89f19a66e79abde550429baec07"} Dec 06 05:42:49 crc kubenswrapper[4957]: I1206 05:42:49.604686 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e62592-5ba0-42ee-b9a0-41bfa6173821","Type":"ContainerStarted","Data":"352f630776e245ced2a350b420554aace0f5575bc1c415f637e232dcc2096351"} Dec 06 05:42:49 crc kubenswrapper[4957]: I1206 05:42:49.615519 4957 generic.go:334] "Generic (PLEG): container finished" podID="fb54e52a-e54e-4d2a-9431-872769baf100" containerID="1c8f296e927917fea00c94076606c536bf420c650c1b55800aa4bd6824209413" exitCode=0 Dec 06 05:42:49 crc kubenswrapper[4957]: I1206 05:42:49.615575 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb54e52a-e54e-4d2a-9431-872769baf100","Type":"ContainerDied","Data":"1c8f296e927917fea00c94076606c536bf420c650c1b55800aa4bd6824209413"} Dec 06 05:42:49 crc kubenswrapper[4957]: I1206 05:42:49.622212 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-slqnh" podStartSLOduration=191.622183924 podStartE2EDuration="3m11.622183924s" podCreationTimestamp="2025-12-06 05:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:42:49.621016542 +0000 UTC m=+209.271284194" watchObservedRunningTime="2025-12-06 05:42:49.622183924 +0000 UTC m=+209.272451586" Dec 06 05:42:49 crc kubenswrapper[4957]: I1206 05:42:49.669526 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=19.669510328 podStartE2EDuration="19.669510328s" podCreationTimestamp="2025-12-06 05:42:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:42:49.649941192 +0000 UTC m=+209.300208834" watchObservedRunningTime="2025-12-06 05:42:49.669510328 +0000 UTC m=+209.319777960" Dec 06 05:42:50 crc kubenswrapper[4957]: I1206 05:42:50.387240 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vs6qp"] Dec 06 05:42:50 crc kubenswrapper[4957]: I1206 05:42:50.850648 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.051991 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb54e52a-e54e-4d2a-9431-872769baf100-kubelet-dir\") pod \"fb54e52a-e54e-4d2a-9431-872769baf100\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.052045 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb54e52a-e54e-4d2a-9431-872769baf100-kube-api-access\") pod \"fb54e52a-e54e-4d2a-9431-872769baf100\" (UID: \"fb54e52a-e54e-4d2a-9431-872769baf100\") " Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.052896 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb54e52a-e54e-4d2a-9431-872769baf100-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fb54e52a-e54e-4d2a-9431-872769baf100" (UID: "fb54e52a-e54e-4d2a-9431-872769baf100"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.061257 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb54e52a-e54e-4d2a-9431-872769baf100-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fb54e52a-e54e-4d2a-9431-872769baf100" (UID: "fb54e52a-e54e-4d2a-9431-872769baf100"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.153337 4957 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fb54e52a-e54e-4d2a-9431-872769baf100-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.153372 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb54e52a-e54e-4d2a-9431-872769baf100-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.625868 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fb54e52a-e54e-4d2a-9431-872769baf100","Type":"ContainerDied","Data":"dbf8ed06a0a149a8cee5fcc518997aa73d2ee172ad6308977286bc3748dee734"} Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.625913 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbf8ed06a0a149a8cee5fcc518997aa73d2ee172ad6308977286bc3748dee734" Dec 06 05:42:51 crc kubenswrapper[4957]: I1206 05:42:51.625914 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 05:42:57 crc kubenswrapper[4957]: I1206 05:42:57.655355 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44rx8" event={"ID":"0dc111dd-bc98-4dfc-8130-85c33d2271b5","Type":"ContainerStarted","Data":"ebf7b72fc6cedc0f3612c54cdeff84fb8e918f42b664f625e5d0ed1e473ad864"} Dec 06 05:42:57 crc kubenswrapper[4957]: I1206 05:42:57.664928 4957 generic.go:334] "Generic (PLEG): container finished" podID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerID="fe2de32d3a062005f8e807692ecd618f813d43790719dfecdcd4ee94d72f8e20" exitCode=0 Dec 06 05:42:57 crc kubenswrapper[4957]: I1206 05:42:57.664966 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbk94" event={"ID":"5de294a0-fcab-43a3-8001-c32c3965f3f6","Type":"ContainerDied","Data":"fe2de32d3a062005f8e807692ecd618f813d43790719dfecdcd4ee94d72f8e20"} Dec 06 05:42:57 crc kubenswrapper[4957]: I1206 05:42:57.672026 4957 generic.go:334] "Generic (PLEG): container finished" podID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerID="d24ce9a9584511200ea542b47e2b283de7e53ca54f0cff5ce91551212750ff3e" exitCode=0 Dec 06 05:42:57 crc kubenswrapper[4957]: I1206 05:42:57.672077 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hj4gr" event={"ID":"87e7bb85-760f-4c78-99b8-6a8b38e8f70f","Type":"ContainerDied","Data":"d24ce9a9584511200ea542b47e2b283de7e53ca54f0cff5ce91551212750ff3e"} Dec 06 05:42:58 crc kubenswrapper[4957]: I1206 05:42:58.687193 4957 generic.go:334] "Generic (PLEG): container finished" podID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerID="ebf7b72fc6cedc0f3612c54cdeff84fb8e918f42b664f625e5d0ed1e473ad864" exitCode=0 Dec 06 05:42:58 crc kubenswrapper[4957]: I1206 05:42:58.687232 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44rx8" event={"ID":"0dc111dd-bc98-4dfc-8130-85c33d2271b5","Type":"ContainerDied","Data":"ebf7b72fc6cedc0f3612c54cdeff84fb8e918f42b664f625e5d0ed1e473ad864"} Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.693732 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44rx8" event={"ID":"0dc111dd-bc98-4dfc-8130-85c33d2271b5","Type":"ContainerStarted","Data":"b22629efd03338ca3b44ce4baab025ae3291b9cce02a093fcb29576f25a525c9"} Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.695548 4957 generic.go:334] "Generic (PLEG): container finished" podID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerID="94336eb85ebac1c144be1e19606b1b7792113501120787462d880dc6d4f1bbba" exitCode=0 Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.695610 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glf2z" event={"ID":"88ca26d1-ec5d-44ee-9bae-db531711bd72","Type":"ContainerDied","Data":"94336eb85ebac1c144be1e19606b1b7792113501120787462d880dc6d4f1bbba"} Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.699470 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbk94" event={"ID":"5de294a0-fcab-43a3-8001-c32c3965f3f6","Type":"ContainerStarted","Data":"93ba862f38abdd563c612d142e371238f4ba4e805eb9cc42e30f51497e9e1181"} Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.703535 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hj4gr" event={"ID":"87e7bb85-760f-4c78-99b8-6a8b38e8f70f","Type":"ContainerStarted","Data":"1b9cc0fb0f4c20a4b4d6795c4099fa80416d82528d8d2b5d6857018802b0f973"} Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.725255 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-44rx8" podStartSLOduration=3.64619019 podStartE2EDuration="1m13.725219768s" podCreationTimestamp="2025-12-06 05:41:46 +0000 UTC" firstStartedPulling="2025-12-06 05:41:49.039653553 +0000 UTC m=+148.689921185" lastFinishedPulling="2025-12-06 05:42:59.118683131 +0000 UTC m=+218.768950763" observedRunningTime="2025-12-06 05:42:59.715980205 +0000 UTC m=+219.366247857" watchObservedRunningTime="2025-12-06 05:42:59.725219768 +0000 UTC m=+219.375487410" Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.736393 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hj4gr" podStartSLOduration=4.121816918 podStartE2EDuration="1m13.736373955s" podCreationTimestamp="2025-12-06 05:41:46 +0000 UTC" firstStartedPulling="2025-12-06 05:41:49.0586995 +0000 UTC m=+148.708967132" lastFinishedPulling="2025-12-06 05:42:58.673256537 +0000 UTC m=+218.323524169" observedRunningTime="2025-12-06 05:42:59.73341612 +0000 UTC m=+219.383683782" watchObservedRunningTime="2025-12-06 05:42:59.736373955 +0000 UTC m=+219.386641587" Dec 06 05:42:59 crc kubenswrapper[4957]: I1206 05:42:59.751120 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kbk94" podStartSLOduration=4.293484432 podStartE2EDuration="1m13.751098313s" podCreationTimestamp="2025-12-06 05:41:46 +0000 UTC" firstStartedPulling="2025-12-06 05:41:49.055180301 +0000 UTC m=+148.705447933" lastFinishedPulling="2025-12-06 05:42:58.512794172 +0000 UTC m=+218.163061814" observedRunningTime="2025-12-06 05:42:59.749235249 +0000 UTC m=+219.399502901" watchObservedRunningTime="2025-12-06 05:42:59.751098313 +0000 UTC m=+219.401365945" Dec 06 05:43:00 crc kubenswrapper[4957]: I1206 05:43:00.711594 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzt5s" event={"ID":"f6be4201-af3d-4f37-a819-7cf136f1e906","Type":"ContainerStarted","Data":"2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8"} Dec 06 05:43:00 crc kubenswrapper[4957]: I1206 05:43:00.714563 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glf2z" event={"ID":"88ca26d1-ec5d-44ee-9bae-db531711bd72","Type":"ContainerStarted","Data":"0c575c01c3a70c5ac7e74ad250d4fad45693a81ef99c329541f89de52d88e316"} Dec 06 05:43:01 crc kubenswrapper[4957]: I1206 05:43:01.683695 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-glf2z" podStartSLOduration=3.671564792 podStartE2EDuration="1m14.683672831s" podCreationTimestamp="2025-12-06 05:41:47 +0000 UTC" firstStartedPulling="2025-12-06 05:41:49.062492496 +0000 UTC m=+148.712760128" lastFinishedPulling="2025-12-06 05:43:00.074600535 +0000 UTC m=+219.724868167" observedRunningTime="2025-12-06 05:43:00.756144382 +0000 UTC m=+220.406412024" watchObservedRunningTime="2025-12-06 05:43:01.683672831 +0000 UTC m=+221.333940483" Dec 06 05:43:01 crc kubenswrapper[4957]: I1206 05:43:01.723369 4957 generic.go:334] "Generic (PLEG): container finished" podID="37639dbf-8f08-428e-93e2-456e2946333f" containerID="43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912" exitCode=0 Dec 06 05:43:01 crc kubenswrapper[4957]: I1206 05:43:01.723433 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2f5bt" event={"ID":"37639dbf-8f08-428e-93e2-456e2946333f","Type":"ContainerDied","Data":"43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912"} Dec 06 05:43:01 crc kubenswrapper[4957]: I1206 05:43:01.727722 4957 generic.go:334] "Generic (PLEG): container finished" podID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerID="3c32cc9b473a7eb9919df12704e4bc84d2c03e40702c1d25b4f470585673c806" exitCode=0 Dec 06 05:43:01 crc kubenswrapper[4957]: I1206 05:43:01.727803 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5chm" event={"ID":"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4","Type":"ContainerDied","Data":"3c32cc9b473a7eb9919df12704e4bc84d2c03e40702c1d25b4f470585673c806"} Dec 06 05:43:01 crc kubenswrapper[4957]: I1206 05:43:01.729880 4957 generic.go:334] "Generic (PLEG): container finished" podID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerID="2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8" exitCode=0 Dec 06 05:43:01 crc kubenswrapper[4957]: I1206 05:43:01.729905 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzt5s" event={"ID":"f6be4201-af3d-4f37-a819-7cf136f1e906","Type":"ContainerDied","Data":"2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8"} Dec 06 05:43:05 crc kubenswrapper[4957]: I1206 05:43:05.759764 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2f5bt" event={"ID":"37639dbf-8f08-428e-93e2-456e2946333f","Type":"ContainerStarted","Data":"654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d"} Dec 06 05:43:05 crc kubenswrapper[4957]: I1206 05:43:05.784209 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2f5bt" podStartSLOduration=5.641177441 podStartE2EDuration="1m17.784183416s" podCreationTimestamp="2025-12-06 05:41:48 +0000 UTC" firstStartedPulling="2025-12-06 05:41:51.093564716 +0000 UTC m=+150.743832358" lastFinishedPulling="2025-12-06 05:43:03.236570701 +0000 UTC m=+222.886838333" observedRunningTime="2025-12-06 05:43:05.777995411 +0000 UTC m=+225.428263043" watchObservedRunningTime="2025-12-06 05:43:05.784183416 +0000 UTC m=+225.434451078" Dec 06 05:43:06 crc kubenswrapper[4957]: I1206 05:43:06.794448 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:43:06 crc kubenswrapper[4957]: I1206 05:43:06.795015 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:43:06 crc kubenswrapper[4957]: I1206 05:43:06.928019 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.026803 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.026984 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.092066 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.233963 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.234101 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.278304 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.449014 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.449621 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.516673 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.839393 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.840685 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.840738 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:43:07 crc kubenswrapper[4957]: I1206 05:43:07.843949 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:43:08 crc kubenswrapper[4957]: I1206 05:43:08.784218 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5chm" event={"ID":"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4","Type":"ContainerStarted","Data":"2eb4e36ea606bfb276581a1e7e40d7cacf2bd2c4af2f7b394b35e90aabfee1ac"} Dec 06 05:43:08 crc kubenswrapper[4957]: I1206 05:43:08.803097 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g5chm" podStartSLOduration=3.842580948 podStartE2EDuration="1m19.803068591s" podCreationTimestamp="2025-12-06 05:41:49 +0000 UTC" firstStartedPulling="2025-12-06 05:41:51.121081072 +0000 UTC m=+150.771348704" lastFinishedPulling="2025-12-06 05:43:07.081568715 +0000 UTC m=+226.731836347" observedRunningTime="2025-12-06 05:43:08.80234006 +0000 UTC m=+228.452607702" watchObservedRunningTime="2025-12-06 05:43:08.803068591 +0000 UTC m=+228.453336223" Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.180982 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.181046 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.237908 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.307365 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hj4gr"] Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.602370 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.602480 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.788866 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hj4gr" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="registry-server" containerID="cri-o://1b9cc0fb0f4c20a4b4d6795c4099fa80416d82528d8d2b5d6857018802b0f973" gracePeriod=2 Dec 06 05:43:09 crc kubenswrapper[4957]: I1206 05:43:09.834115 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:43:10 crc kubenswrapper[4957]: I1206 05:43:10.652585 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-g5chm" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="registry-server" probeResult="failure" output=< Dec 06 05:43:10 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 05:43:10 crc kubenswrapper[4957]: > Dec 06 05:43:10 crc kubenswrapper[4957]: I1206 05:43:10.710161 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glf2z"] Dec 06 05:43:10 crc kubenswrapper[4957]: I1206 05:43:10.710502 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-glf2z" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="registry-server" containerID="cri-o://0c575c01c3a70c5ac7e74ad250d4fad45693a81ef99c329541f89de52d88e316" gracePeriod=2 Dec 06 05:43:13 crc kubenswrapper[4957]: I1206 05:43:13.822110 4957 generic.go:334] "Generic (PLEG): container finished" podID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerID="0c575c01c3a70c5ac7e74ad250d4fad45693a81ef99c329541f89de52d88e316" exitCode=0 Dec 06 05:43:13 crc kubenswrapper[4957]: I1206 05:43:13.822192 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glf2z" event={"ID":"88ca26d1-ec5d-44ee-9bae-db531711bd72","Type":"ContainerDied","Data":"0c575c01c3a70c5ac7e74ad250d4fad45693a81ef99c329541f89de52d88e316"} Dec 06 05:43:13 crc kubenswrapper[4957]: I1206 05:43:13.826302 4957 generic.go:334] "Generic (PLEG): container finished" podID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerID="1b9cc0fb0f4c20a4b4d6795c4099fa80416d82528d8d2b5d6857018802b0f973" exitCode=0 Dec 06 05:43:13 crc kubenswrapper[4957]: I1206 05:43:13.826354 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hj4gr" event={"ID":"87e7bb85-760f-4c78-99b8-6a8b38e8f70f","Type":"ContainerDied","Data":"1b9cc0fb0f4c20a4b4d6795c4099fa80416d82528d8d2b5d6857018802b0f973"} Dec 06 05:43:14 crc kubenswrapper[4957]: I1206 05:43:14.834661 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glf2z" event={"ID":"88ca26d1-ec5d-44ee-9bae-db531711bd72","Type":"ContainerDied","Data":"de34137f3316b417536fea72799f77996cc2a2eec4b732f12f7562e4de16086a"} Dec 06 05:43:14 crc kubenswrapper[4957]: I1206 05:43:14.834745 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de34137f3316b417536fea72799f77996cc2a2eec4b732f12f7562e4de16086a" Dec 06 05:43:14 crc kubenswrapper[4957]: I1206 05:43:14.839348 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hj4gr" event={"ID":"87e7bb85-760f-4c78-99b8-6a8b38e8f70f","Type":"ContainerDied","Data":"cf55a42b23539cfad5d1b63acd31d43ced8702c31830c6cf37a8f444e3701e19"} Dec 06 05:43:14 crc kubenswrapper[4957]: I1206 05:43:14.839398 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf55a42b23539cfad5d1b63acd31d43ced8702c31830c6cf37a8f444e3701e19" Dec 06 05:43:14 crc kubenswrapper[4957]: I1206 05:43:14.851756 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:43:14 crc kubenswrapper[4957]: I1206 05:43:14.858211 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.040237 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfkfz\" (UniqueName: \"kubernetes.io/projected/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-kube-api-access-gfkfz\") pod \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.040300 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtqzx\" (UniqueName: \"kubernetes.io/projected/88ca26d1-ec5d-44ee-9bae-db531711bd72-kube-api-access-xtqzx\") pod \"88ca26d1-ec5d-44ee-9bae-db531711bd72\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.040398 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-utilities\") pod \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.040438 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-catalog-content\") pod \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\" (UID: \"87e7bb85-760f-4c78-99b8-6a8b38e8f70f\") " Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.040483 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-catalog-content\") pod \"88ca26d1-ec5d-44ee-9bae-db531711bd72\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.040563 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-utilities\") pod \"88ca26d1-ec5d-44ee-9bae-db531711bd72\" (UID: \"88ca26d1-ec5d-44ee-9bae-db531711bd72\") " Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.042132 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-utilities" (OuterVolumeSpecName: "utilities") pod "88ca26d1-ec5d-44ee-9bae-db531711bd72" (UID: "88ca26d1-ec5d-44ee-9bae-db531711bd72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.042210 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-utilities" (OuterVolumeSpecName: "utilities") pod "87e7bb85-760f-4c78-99b8-6a8b38e8f70f" (UID: "87e7bb85-760f-4c78-99b8-6a8b38e8f70f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.049035 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ca26d1-ec5d-44ee-9bae-db531711bd72-kube-api-access-xtqzx" (OuterVolumeSpecName: "kube-api-access-xtqzx") pod "88ca26d1-ec5d-44ee-9bae-db531711bd72" (UID: "88ca26d1-ec5d-44ee-9bae-db531711bd72"). InnerVolumeSpecName "kube-api-access-xtqzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.049182 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-kube-api-access-gfkfz" (OuterVolumeSpecName: "kube-api-access-gfkfz") pod "87e7bb85-760f-4c78-99b8-6a8b38e8f70f" (UID: "87e7bb85-760f-4c78-99b8-6a8b38e8f70f"). InnerVolumeSpecName "kube-api-access-gfkfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.096932 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87e7bb85-760f-4c78-99b8-6a8b38e8f70f" (UID: "87e7bb85-760f-4c78-99b8-6a8b38e8f70f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.142981 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.143025 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.143036 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfkfz\" (UniqueName: \"kubernetes.io/projected/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-kube-api-access-gfkfz\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.143050 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtqzx\" (UniqueName: \"kubernetes.io/projected/88ca26d1-ec5d-44ee-9bae-db531711bd72-kube-api-access-xtqzx\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.143060 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87e7bb85-760f-4c78-99b8-6a8b38e8f70f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.147770 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88ca26d1-ec5d-44ee-9bae-db531711bd72" (UID: "88ca26d1-ec5d-44ee-9bae-db531711bd72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.244906 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88ca26d1-ec5d-44ee-9bae-db531711bd72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.418007 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" podUID="70c2a386-5150-4abb-9e97-af87740cc325" containerName="oauth-openshift" containerID="cri-o://90640a5a65d1169b1443983bce4931a15c506353e8de3c5f88b05aef67b07cb4" gracePeriod=15 Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.845612 4957 generic.go:334] "Generic (PLEG): container finished" podID="70c2a386-5150-4abb-9e97-af87740cc325" containerID="90640a5a65d1169b1443983bce4931a15c506353e8de3c5f88b05aef67b07cb4" exitCode=0 Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.845735 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" event={"ID":"70c2a386-5150-4abb-9e97-af87740cc325","Type":"ContainerDied","Data":"90640a5a65d1169b1443983bce4931a15c506353e8de3c5f88b05aef67b07cb4"} Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.846010 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hj4gr" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.846693 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glf2z" Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.883939 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hj4gr"] Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.887166 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hj4gr"] Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.898125 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glf2z"] Dec 06 05:43:15 crc kubenswrapper[4957]: I1206 05:43:15.901581 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-glf2z"] Dec 06 05:43:16 crc kubenswrapper[4957]: I1206 05:43:16.377251 4957 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vs6qp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" start-of-body= Dec 06 05:43:16 crc kubenswrapper[4957]: I1206 05:43:16.377349 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" podUID="70c2a386-5150-4abb-9e97-af87740cc325" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" Dec 06 05:43:16 crc kubenswrapper[4957]: I1206 05:43:16.669276 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" path="/var/lib/kubelet/pods/87e7bb85-760f-4c78-99b8-6a8b38e8f70f/volumes" Dec 06 05:43:16 crc kubenswrapper[4957]: I1206 05:43:16.670353 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" path="/var/lib/kubelet/pods/88ca26d1-ec5d-44ee-9bae-db531711bd72/volumes" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.363412 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481517 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-error\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481614 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-service-ca\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481652 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-cliconfig\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481671 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70c2a386-5150-4abb-9e97-af87740cc325-audit-dir\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481700 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-router-certs\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481722 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-idp-0-file-data\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481767 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-audit-policies\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481819 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf99g\" (UniqueName: \"kubernetes.io/projected/70c2a386-5150-4abb-9e97-af87740cc325-kube-api-access-sf99g\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481872 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-serving-cert\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481892 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-session\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481930 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-provider-selection\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.481957 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-ocp-branding-template\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.482001 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-trusted-ca-bundle\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.482034 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-login\") pod \"70c2a386-5150-4abb-9e97-af87740cc325\" (UID: \"70c2a386-5150-4abb-9e97-af87740cc325\") " Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.483304 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70c2a386-5150-4abb-9e97-af87740cc325-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.483910 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.484176 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.484735 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.485643 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.490261 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.490694 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c2a386-5150-4abb-9e97-af87740cc325-kube-api-access-sf99g" (OuterVolumeSpecName: "kube-api-access-sf99g") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "kube-api-access-sf99g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.490757 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.491055 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.492086 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.492202 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.492425 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.492870 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.493072 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "70c2a386-5150-4abb-9e97-af87740cc325" (UID: "70c2a386-5150-4abb-9e97-af87740cc325"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583151 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583197 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583209 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583219 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583230 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583241 4957 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70c2a386-5150-4abb-9e97-af87740cc325-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583252 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583261 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583270 4957 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/70c2a386-5150-4abb-9e97-af87740cc325-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583279 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf99g\" (UniqueName: \"kubernetes.io/projected/70c2a386-5150-4abb-9e97-af87740cc325-kube-api-access-sf99g\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583290 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583298 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583307 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.583317 4957 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/70c2a386-5150-4abb-9e97-af87740cc325-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.864682 4957 generic.go:334] "Generic (PLEG): container finished" podID="b5b5df82-a892-4971-84e3-3c3690960f71" containerID="e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8" exitCode=0 Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.864771 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hz4nt" event={"ID":"b5b5df82-a892-4971-84e3-3c3690960f71","Type":"ContainerDied","Data":"e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8"} Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.870763 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzt5s" event={"ID":"f6be4201-af3d-4f37-a819-7cf136f1e906","Type":"ContainerStarted","Data":"6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf"} Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.887477 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" event={"ID":"70c2a386-5150-4abb-9e97-af87740cc325","Type":"ContainerDied","Data":"ebcca660a2754e891cfb8e89967c8a2eebb94238757aae4120009f0af5516bd5"} Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.887532 4957 scope.go:117] "RemoveContainer" containerID="90640a5a65d1169b1443983bce4931a15c506353e8de3c5f88b05aef67b07cb4" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.887669 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vs6qp" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.924418 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mzt5s" podStartSLOduration=4.648603224 podStartE2EDuration="1m28.924397508s" podCreationTimestamp="2025-12-06 05:41:49 +0000 UTC" firstStartedPulling="2025-12-06 05:41:51.137302349 +0000 UTC m=+150.787569981" lastFinishedPulling="2025-12-06 05:43:15.413096633 +0000 UTC m=+235.063364265" observedRunningTime="2025-12-06 05:43:17.920042304 +0000 UTC m=+237.570309946" watchObservedRunningTime="2025-12-06 05:43:17.924397508 +0000 UTC m=+237.574665150" Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.941954 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vs6qp"] Dec 06 05:43:17 crc kubenswrapper[4957]: I1206 05:43:17.948184 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vs6qp"] Dec 06 05:43:18 crc kubenswrapper[4957]: I1206 05:43:18.671186 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c2a386-5150-4abb-9e97-af87740cc325" path="/var/lib/kubelet/pods/70c2a386-5150-4abb-9e97-af87740cc325/volumes" Dec 06 05:43:18 crc kubenswrapper[4957]: I1206 05:43:18.906685 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hz4nt" event={"ID":"b5b5df82-a892-4971-84e3-3c3690960f71","Type":"ContainerStarted","Data":"31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762"} Dec 06 05:43:18 crc kubenswrapper[4957]: I1206 05:43:18.927929 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hz4nt" podStartSLOduration=2.780504709 podStartE2EDuration="1m29.927896533s" podCreationTimestamp="2025-12-06 05:41:49 +0000 UTC" firstStartedPulling="2025-12-06 05:41:51.126147595 +0000 UTC m=+150.776415227" lastFinishedPulling="2025-12-06 05:43:18.273539389 +0000 UTC m=+237.923807051" observedRunningTime="2025-12-06 05:43:18.922658374 +0000 UTC m=+238.572926036" watchObservedRunningTime="2025-12-06 05:43:18.927896533 +0000 UTC m=+238.578164185" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.229940 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-55c7db9594-lxc8c"] Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230276 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="extract-content" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230295 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="extract-content" Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230320 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="extract-utilities" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230329 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="extract-utilities" Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230339 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="registry-server" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230349 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="registry-server" Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230360 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb54e52a-e54e-4d2a-9431-872769baf100" containerName="pruner" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230368 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb54e52a-e54e-4d2a-9431-872769baf100" containerName="pruner" Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230384 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="extract-utilities" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230398 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="extract-utilities" Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230411 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c2a386-5150-4abb-9e97-af87740cc325" containerName="oauth-openshift" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230419 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c2a386-5150-4abb-9e97-af87740cc325" containerName="oauth-openshift" Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230434 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="extract-content" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230442 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="extract-content" Dec 06 05:43:19 crc kubenswrapper[4957]: E1206 05:43:19.230455 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="registry-server" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230463 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="registry-server" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230632 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ca26d1-ec5d-44ee-9bae-db531711bd72" containerName="registry-server" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230673 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb54e52a-e54e-4d2a-9431-872769baf100" containerName="pruner" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230687 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e7bb85-760f-4c78-99b8-6a8b38e8f70f" containerName="registry-server" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.230707 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c2a386-5150-4abb-9e97-af87740cc325" containerName="oauth-openshift" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.231423 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.236883 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.237205 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.237666 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.238944 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.239137 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.239648 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.239704 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.239783 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.239949 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.240100 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.240294 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.249819 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55c7db9594-lxc8c"] Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.250267 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.250409 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.253561 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.260416 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.407474 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408041 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-audit-policies\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408188 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-login\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408240 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408339 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5748b\" (UniqueName: \"kubernetes.io/projected/e9cf335b-084f-4556-a7ce-823930880b96-kube-api-access-5748b\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408409 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408469 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408500 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-router-certs\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408526 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-error\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408552 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-service-ca\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408578 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e9cf335b-084f-4556-a7ce-823930880b96-audit-dir\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408600 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408624 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.408656 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-session\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509679 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5748b\" (UniqueName: \"kubernetes.io/projected/e9cf335b-084f-4556-a7ce-823930880b96-kube-api-access-5748b\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509755 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509814 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509867 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-router-certs\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509893 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-error\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509919 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-service-ca\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509945 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e9cf335b-084f-4556-a7ce-823930880b96-audit-dir\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509967 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.509993 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.510023 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-session\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.510046 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.510076 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-audit-policies\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.510101 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-login\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.510126 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.510512 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e9cf335b-084f-4556-a7ce-823930880b96-audit-dir\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.511551 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.512070 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.513632 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-audit-policies\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.514655 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-service-ca\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.516496 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-error\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.516816 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-session\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.517258 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.522287 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.524569 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-login\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.527257 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.533030 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-system-router-certs\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.533108 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e9cf335b-084f-4556-a7ce-823930880b96-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.539941 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5748b\" (UniqueName: \"kubernetes.io/projected/e9cf335b-084f-4556-a7ce-823930880b96-kube-api-access-5748b\") pod \"oauth-openshift-55c7db9594-lxc8c\" (UID: \"e9cf335b-084f-4556-a7ce-823930880b96\") " pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.551579 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.644564 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.736432 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.864978 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55c7db9594-lxc8c"] Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.916083 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" event={"ID":"e9cf335b-084f-4556-a7ce-823930880b96","Type":"ContainerStarted","Data":"bca2c33a647d87835c693b91b50171136384a1ad8e9726ff62762edb12736859"} Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.983245 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:43:19 crc kubenswrapper[4957]: I1206 05:43:19.983332 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:43:20 crc kubenswrapper[4957]: I1206 05:43:20.197709 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:43:20 crc kubenswrapper[4957]: I1206 05:43:20.197761 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:43:21 crc kubenswrapper[4957]: I1206 05:43:21.038468 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mzt5s" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="registry-server" probeResult="failure" output=< Dec 06 05:43:21 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 05:43:21 crc kubenswrapper[4957]: > Dec 06 05:43:21 crc kubenswrapper[4957]: I1206 05:43:21.233082 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hz4nt" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="registry-server" probeResult="failure" output=< Dec 06 05:43:21 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 05:43:21 crc kubenswrapper[4957]: > Dec 06 05:43:23 crc kubenswrapper[4957]: I1206 05:43:23.305265 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5chm"] Dec 06 05:43:23 crc kubenswrapper[4957]: I1206 05:43:23.305612 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g5chm" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="registry-server" containerID="cri-o://2eb4e36ea606bfb276581a1e7e40d7cacf2bd2c4af2f7b394b35e90aabfee1ac" gracePeriod=2 Dec 06 05:43:24 crc kubenswrapper[4957]: I1206 05:43:24.945784 4957 generic.go:334] "Generic (PLEG): container finished" podID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerID="2eb4e36ea606bfb276581a1e7e40d7cacf2bd2c4af2f7b394b35e90aabfee1ac" exitCode=0 Dec 06 05:43:24 crc kubenswrapper[4957]: I1206 05:43:24.945856 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5chm" event={"ID":"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4","Type":"ContainerDied","Data":"2eb4e36ea606bfb276581a1e7e40d7cacf2bd2c4af2f7b394b35e90aabfee1ac"} Dec 06 05:43:24 crc kubenswrapper[4957]: I1206 05:43:24.948674 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" event={"ID":"e9cf335b-084f-4556-a7ce-823930880b96","Type":"ContainerStarted","Data":"c2c692053144fd207e074a01c46b054977002eb3f49395351d478ddf05c44a30"} Dec 06 05:43:24 crc kubenswrapper[4957]: I1206 05:43:24.949024 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:24 crc kubenswrapper[4957]: I1206 05:43:24.958314 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" Dec 06 05:43:24 crc kubenswrapper[4957]: I1206 05:43:24.971524 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-55c7db9594-lxc8c" podStartSLOduration=34.971498607 podStartE2EDuration="34.971498607s" podCreationTimestamp="2025-12-06 05:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:43:24.968936754 +0000 UTC m=+244.619204406" watchObservedRunningTime="2025-12-06 05:43:24.971498607 +0000 UTC m=+244.621766239" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.263226 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.389526 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-567zz\" (UniqueName: \"kubernetes.io/projected/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-kube-api-access-567zz\") pod \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.389608 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-utilities\") pod \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.389686 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-catalog-content\") pod \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\" (UID: \"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4\") " Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.391119 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-utilities" (OuterVolumeSpecName: "utilities") pod "e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" (UID: "e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.396263 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-kube-api-access-567zz" (OuterVolumeSpecName: "kube-api-access-567zz") pod "e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" (UID: "e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4"). InnerVolumeSpecName "kube-api-access-567zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.397382 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-567zz\" (UniqueName: \"kubernetes.io/projected/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-kube-api-access-567zz\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.397434 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.421192 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" (UID: "e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.498970 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.956810 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5chm" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.956801 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5chm" event={"ID":"e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4","Type":"ContainerDied","Data":"6ad7804cbab340b5c2d676930e44abf947179031754416ae161f3c70924d2e0b"} Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.956923 4957 scope.go:117] "RemoveContainer" containerID="2eb4e36ea606bfb276581a1e7e40d7cacf2bd2c4af2f7b394b35e90aabfee1ac" Dec 06 05:43:25 crc kubenswrapper[4957]: I1206 05:43:25.989690 4957 scope.go:117] "RemoveContainer" containerID="3c32cc9b473a7eb9919df12704e4bc84d2c03e40702c1d25b4f470585673c806" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.001947 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5chm"] Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.005042 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5chm"] Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.032332 4957 scope.go:117] "RemoveContainer" containerID="7addc9291f4aa344f94265482364f8ed0bf3c79f5631d14c262262e61ec43090" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.377195 4957 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.377511 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d" gracePeriod=15 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.377664 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488" gracePeriod=15 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.377714 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380" gracePeriod=15 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.377742 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c" gracePeriod=15 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.377774 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648" gracePeriod=15 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381073 4957 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381325 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="registry-server" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381346 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="registry-server" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381362 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381370 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381381 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381388 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381405 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381414 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381426 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381433 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381445 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381477 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381488 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381496 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381507 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="extract-content" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381515 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="extract-content" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381526 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381536 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.381549 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="extract-utilities" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381558 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="extract-utilities" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381683 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" containerName="registry-server" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381702 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381713 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381722 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381730 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381739 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.381748 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.383058 4957 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.383681 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.388673 4957 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.429194 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.515159 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.515416 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.515529 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.515613 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.515703 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.515796 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.515910 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.516164 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617031 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617088 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617121 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617154 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617162 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617217 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617247 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617253 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617220 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617278 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617313 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617336 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617367 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617386 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617417 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.617454 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.672561 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4" path="/var/lib/kubelet/pods/e2e65b89-5a7e-4845-9bd7-6ae4c3e936f4/volumes" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.783527 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.790515 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:26Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:26Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:26Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:26Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:544a01170a4aa6cf8322d5bffa5817113efd696e3c3e9bac6a29d2da9f9451e5\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:67f42a86b99b69b357285a6845977f967e6c825de2049c19620a78eaf99cebf3\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1222075732},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:2ebdb257c6ae11dce352087cce880750555ee33cde3006a4607a1305a2bd1b20\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1e1cdc371970de5a5af444ac8a3d7e6bb4086055fac0cc443b1d09c4ad585ab\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201795919},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.790962 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.791293 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.791692 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.791895 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.791920 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:43:26 crc kubenswrapper[4957]: W1206 05:43:26.819008 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-4f720f38b7ffd227eb76fb9cbeb4db1ff27e131e57d8d2966a41cab9e2e48b16 WatchSource:0}: Error finding container 4f720f38b7ffd227eb76fb9cbeb4db1ff27e131e57d8d2966a41cab9e2e48b16: Status 404 returned error can't find the container with id 4f720f38b7ffd227eb76fb9cbeb4db1ff27e131e57d8d2966a41cab9e2e48b16 Dec 06 05:43:26 crc kubenswrapper[4957]: E1206 05:43:26.823733 4957 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.151:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e89f912460dbd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:43:26.822256061 +0000 UTC m=+246.472523693,LastTimestamp:2025-12-06 05:43:26.822256061 +0000 UTC m=+246.472523693,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.969377 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.970910 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.971756 4957 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488" exitCode=0 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.971777 4957 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380" exitCode=0 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.971787 4957 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c" exitCode=0 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.971794 4957 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648" exitCode=2 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.971871 4957 scope.go:117] "RemoveContainer" containerID="1ba6739206a4ed3d5d9f9aef82879494c9cc52751d2fbd2b2a728b537532b9d9" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.973953 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4f720f38b7ffd227eb76fb9cbeb4db1ff27e131e57d8d2966a41cab9e2e48b16"} Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.979507 4957 generic.go:334] "Generic (PLEG): container finished" podID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" containerID="352f630776e245ced2a350b420554aace0f5575bc1c415f637e232dcc2096351" exitCode=0 Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.979613 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e62592-5ba0-42ee-b9a0-41bfa6173821","Type":"ContainerDied","Data":"352f630776e245ced2a350b420554aace0f5575bc1c415f637e232dcc2096351"} Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.980308 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:26 crc kubenswrapper[4957]: I1206 05:43:26.980582 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:27 crc kubenswrapper[4957]: I1206 05:43:27.991517 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 05:43:27 crc kubenswrapper[4957]: I1206 05:43:27.994773 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e"} Dec 06 05:43:27 crc kubenswrapper[4957]: I1206 05:43:27.995891 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:27 crc kubenswrapper[4957]: I1206 05:43:27.996713 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.335327 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.337157 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.337891 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.442691 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kubelet-dir\") pod \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.443601 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-var-lock\") pod \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.442862 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e7e62592-5ba0-42ee-b9a0-41bfa6173821" (UID: "e7e62592-5ba0-42ee-b9a0-41bfa6173821"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.443688 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kube-api-access\") pod \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\" (UID: \"e7e62592-5ba0-42ee-b9a0-41bfa6173821\") " Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.443750 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-var-lock" (OuterVolumeSpecName: "var-lock") pod "e7e62592-5ba0-42ee-b9a0-41bfa6173821" (UID: "e7e62592-5ba0-42ee-b9a0-41bfa6173821"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.444056 4957 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.444086 4957 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e62592-5ba0-42ee-b9a0-41bfa6173821-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.452480 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e62592-5ba0-42ee-b9a0-41bfa6173821" (UID: "e7e62592-5ba0-42ee-b9a0-41bfa6173821"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.546172 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e62592-5ba0-42ee-b9a0-41bfa6173821-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.861000 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.862333 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.863401 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.863801 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.864394 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.950639 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.950746 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.950733 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.950797 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.950808 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.950880 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.951052 4957 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.951071 4957 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:28 crc kubenswrapper[4957]: I1206 05:43:28.951082 4957 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.024713 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e62592-5ba0-42ee-b9a0-41bfa6173821","Type":"ContainerDied","Data":"09b4be6e7ae9d04f91a618f478fccc6074474cc2d7a60b155562bcae6d27b1d7"} Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.024786 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09b4be6e7ae9d04f91a618f478fccc6074474cc2d7a60b155562bcae6d27b1d7" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.024731 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.028161 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.029029 4957 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d" exitCode=0 Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.029128 4957 scope.go:117] "RemoveContainer" containerID="9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.029243 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.029307 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.029739 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.030309 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.049661 4957 scope.go:117] "RemoveContainer" containerID="f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.058592 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.058907 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.059154 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.064958 4957 scope.go:117] "RemoveContainer" containerID="1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.083379 4957 scope.go:117] "RemoveContainer" containerID="155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.099156 4957 scope.go:117] "RemoveContainer" containerID="7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.124213 4957 scope.go:117] "RemoveContainer" containerID="d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.148256 4957 scope.go:117] "RemoveContainer" containerID="9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488" Dec 06 05:43:29 crc kubenswrapper[4957]: E1206 05:43:29.148977 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\": container with ID starting with 9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488 not found: ID does not exist" containerID="9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.149032 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488"} err="failed to get container status \"9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\": rpc error: code = NotFound desc = could not find container \"9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488\": container with ID starting with 9c1bd313badc8aadef84fcfb98dce5419f468ffd77935a6ade1a7754343e7488 not found: ID does not exist" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.149077 4957 scope.go:117] "RemoveContainer" containerID="f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380" Dec 06 05:43:29 crc kubenswrapper[4957]: E1206 05:43:29.149583 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\": container with ID starting with f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380 not found: ID does not exist" containerID="f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.149612 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380"} err="failed to get container status \"f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\": rpc error: code = NotFound desc = could not find container \"f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380\": container with ID starting with f84bf843252cceba8107e31a48f22263024977dc50a5850c54a720dc109f2380 not found: ID does not exist" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.149632 4957 scope.go:117] "RemoveContainer" containerID="1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c" Dec 06 05:43:29 crc kubenswrapper[4957]: E1206 05:43:29.150235 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\": container with ID starting with 1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c not found: ID does not exist" containerID="1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.150254 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c"} err="failed to get container status \"1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\": rpc error: code = NotFound desc = could not find container \"1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c\": container with ID starting with 1b417c9161759364f004cb696fd01612a3a8051d8f55cdd11831a55642a9bb3c not found: ID does not exist" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.150269 4957 scope.go:117] "RemoveContainer" containerID="155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648" Dec 06 05:43:29 crc kubenswrapper[4957]: E1206 05:43:29.150614 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\": container with ID starting with 155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648 not found: ID does not exist" containerID="155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.150630 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648"} err="failed to get container status \"155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\": rpc error: code = NotFound desc = could not find container \"155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648\": container with ID starting with 155b3509afc3a6d297f925de44fb4f544478f83577b057e62789a43161df1648 not found: ID does not exist" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.150655 4957 scope.go:117] "RemoveContainer" containerID="7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d" Dec 06 05:43:29 crc kubenswrapper[4957]: E1206 05:43:29.150926 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\": container with ID starting with 7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d not found: ID does not exist" containerID="7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.150944 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d"} err="failed to get container status \"7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\": rpc error: code = NotFound desc = could not find container \"7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d\": container with ID starting with 7838d0c2ddcf7c267428f4673f81fda7e9c388981f65d4bc0a9fd5671349a55d not found: ID does not exist" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.150957 4957 scope.go:117] "RemoveContainer" containerID="d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f" Dec 06 05:43:29 crc kubenswrapper[4957]: E1206 05:43:29.151233 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\": container with ID starting with d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f not found: ID does not exist" containerID="d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f" Dec 06 05:43:29 crc kubenswrapper[4957]: I1206 05:43:29.151260 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f"} err="failed to get container status \"d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\": rpc error: code = NotFound desc = could not find container \"d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f\": container with ID starting with d0d50ea06635266d61be32e5b26e07d0076dd4d79e3d9d1f86e54aed3562874f not found: ID does not exist" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.034501 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.034947 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.035463 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.035917 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.036193 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.074618 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.075978 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.076994 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.077949 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.078417 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.239170 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.239985 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.240252 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.240607 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.241017 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.241383 4957 status_manager.go:851] "Failed to get status for pod" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" pod="openshift-marketplace/redhat-operators-hz4nt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hz4nt\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.278006 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.278501 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.278886 4957 status_manager.go:851] "Failed to get status for pod" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" pod="openshift-marketplace/redhat-operators-hz4nt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hz4nt\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.279397 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.279632 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.279924 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.667146 4957 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.667878 4957 status_manager.go:851] "Failed to get status for pod" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" pod="openshift-marketplace/redhat-operators-hz4nt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hz4nt\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.668289 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.668882 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.669530 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:30 crc kubenswrapper[4957]: I1206 05:43:30.671008 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 05:43:31 crc kubenswrapper[4957]: E1206 05:43:31.849287 4957 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.151:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e89f912460dbd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 05:43:26.822256061 +0000 UTC m=+246.472523693,LastTimestamp:2025-12-06 05:43:26.822256061 +0000 UTC m=+246.472523693,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.068661 4957 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.069682 4957 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.070266 4957 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.070707 4957 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.071131 4957 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: I1206 05:43:36.071172 4957 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.071476 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="200ms" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.272780 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="400ms" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.673927 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="800ms" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.880459 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:36Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:36Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:36Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T05:43:36Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:544a01170a4aa6cf8322d5bffa5817113efd696e3c3e9bac6a29d2da9f9451e5\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:67f42a86b99b69b357285a6845977f967e6c825de2049c19620a78eaf99cebf3\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1222075732},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:2ebdb257c6ae11dce352087cce880750555ee33cde3006a4607a1305a2bd1b20\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1e1cdc371970de5a5af444ac8a3d7e6bb4086055fac0cc443b1d09c4ad585ab\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201795919},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.881444 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.882004 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.882392 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.882934 4957 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:36 crc kubenswrapper[4957]: E1206 05:43:36.882966 4957 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 05:43:37 crc kubenswrapper[4957]: E1206 05:43:37.474814 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="1.6s" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.661870 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.662425 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.662750 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.663408 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.663730 4957 status_manager.go:851] "Failed to get status for pod" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" pod="openshift-marketplace/redhat-operators-hz4nt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hz4nt\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.675214 4957 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.675251 4957 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:38 crc kubenswrapper[4957]: E1206 05:43:38.675706 4957 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:38 crc kubenswrapper[4957]: I1206 05:43:38.676149 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:39 crc kubenswrapper[4957]: E1206 05:43:39.075597 4957 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.151:6443: connect: connection refused" interval="3.2s" Dec 06 05:43:39 crc kubenswrapper[4957]: I1206 05:43:39.101495 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f1173a968fb9b754656d8f42b316e07c8ab686f9d12b3e94d1ea03f6c2d7aa50"} Dec 06 05:43:40 crc kubenswrapper[4957]: I1206 05:43:40.673149 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:40 crc kubenswrapper[4957]: I1206 05:43:40.673866 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:40 crc kubenswrapper[4957]: I1206 05:43:40.674173 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:40 crc kubenswrapper[4957]: I1206 05:43:40.675044 4957 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:40 crc kubenswrapper[4957]: I1206 05:43:40.676072 4957 status_manager.go:851] "Failed to get status for pod" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" pod="openshift-marketplace/redhat-operators-hz4nt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hz4nt\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.118608 4957 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b818bb5db0328b3bd0601388378098177112036255f96cee4a2a1f71f8ed0bba" exitCode=0 Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.118744 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b818bb5db0328b3bd0601388378098177112036255f96cee4a2a1f71f8ed0bba"} Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.119194 4957 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.119234 4957 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.119417 4957 status_manager.go:851] "Failed to get status for pod" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.119667 4957 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:41 crc kubenswrapper[4957]: E1206 05:43:41.119798 4957 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.119929 4957 status_manager.go:851] "Failed to get status for pod" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" pod="openshift-marketplace/redhat-operators-hz4nt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hz4nt\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.120125 4957 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.120310 4957 status_manager.go:851] "Failed to get status for pod" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" pod="openshift-marketplace/redhat-operators-mzt5s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzt5s\": dial tcp 38.102.83.151:6443: connect: connection refused" Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.383054 4957 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": read tcp 192.168.126.11:33688->192.168.126.11:10257: read: connection reset by peer" start-of-body= Dec 06 05:43:41 crc kubenswrapper[4957]: I1206 05:43:41.383110 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": read tcp 192.168.126.11:33688->192.168.126.11:10257: read: connection reset by peer" Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.135027 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.137513 4957 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd" exitCode=1 Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.137610 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd"} Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.147411 4957 scope.go:117] "RemoveContainer" containerID="e5d40bb6e9bb931fb3695579816c55395b0d2d030704bfb664501fdc337ed8bd" Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.152553 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4a68bd71c4eb095de533c1978644b5e0dc2b0eced9441851d906eea7216db532"} Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.152609 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c9be989073ed532efaafb2e0eb5e2c887034c1db01c02dcb0cc04a32f6845df7"} Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.152623 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5a4fa8e68108a954dcc455d75ade6b6c7c847b0950d055abf827638080fbb699"} Dec 06 05:43:42 crc kubenswrapper[4957]: I1206 05:43:42.152639 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"073fa582e3df4dd054d753b2b39f3dfebc7c8c606247b9f43a531b35a30298be"} Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.162497 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.162617 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ac93ccb61de020af7106439e4699f91de4d08ae8305685ec47ef23918604d77b"} Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.167464 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d18ff9eb2c30dd436a76b2d2a270d704e722acb710c59c6e0b52f887d38235b7"} Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.167724 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.167824 4957 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.167875 4957 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.676535 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.677041 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.681288 4957 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]log ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]etcd ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/generic-apiserver-start-informers ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/priority-and-fairness-filter ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-apiextensions-informers ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-apiextensions-controllers ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/crd-informer-synced ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-system-namespaces-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 06 05:43:43 crc kubenswrapper[4957]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 06 05:43:43 crc kubenswrapper[4957]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/bootstrap-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/start-kube-aggregator-informers ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/apiservice-registration-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/apiservice-discovery-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]autoregister-completion ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/apiservice-openapi-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 06 05:43:43 crc kubenswrapper[4957]: livez check failed Dec 06 05:43:43 crc kubenswrapper[4957]: I1206 05:43:43.681352 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 05:43:48 crc kubenswrapper[4957]: I1206 05:43:48.176902 4957 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:48 crc kubenswrapper[4957]: I1206 05:43:48.683329 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:48 crc kubenswrapper[4957]: I1206 05:43:48.686550 4957 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="feae7ae1-0a1f-4aa6-bdd4-32d629f14f8d" Dec 06 05:43:49 crc kubenswrapper[4957]: I1206 05:43:49.218989 4957 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:49 crc kubenswrapper[4957]: I1206 05:43:49.219362 4957 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:49 crc kubenswrapper[4957]: I1206 05:43:49.762675 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:43:49 crc kubenswrapper[4957]: I1206 05:43:49.774253 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:43:50 crc kubenswrapper[4957]: I1206 05:43:50.223683 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:43:50 crc kubenswrapper[4957]: I1206 05:43:50.223866 4957 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:50 crc kubenswrapper[4957]: I1206 05:43:50.223895 4957 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:50 crc kubenswrapper[4957]: I1206 05:43:50.228551 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:43:50 crc kubenswrapper[4957]: I1206 05:43:50.689907 4957 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="feae7ae1-0a1f-4aa6-bdd4-32d629f14f8d" Dec 06 05:43:51 crc kubenswrapper[4957]: I1206 05:43:51.228283 4957 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:51 crc kubenswrapper[4957]: I1206 05:43:51.228323 4957 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e395b2fa-2c3f-4dfd-bdb1-9c7969c3d39e" Dec 06 05:43:51 crc kubenswrapper[4957]: I1206 05:43:51.231333 4957 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="feae7ae1-0a1f-4aa6-bdd4-32d629f14f8d" Dec 06 05:43:57 crc kubenswrapper[4957]: I1206 05:43:57.833163 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 05:43:58 crc kubenswrapper[4957]: I1206 05:43:58.256670 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 05:43:58 crc kubenswrapper[4957]: I1206 05:43:58.328556 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 05:43:58 crc kubenswrapper[4957]: I1206 05:43:58.382726 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 05:43:58 crc kubenswrapper[4957]: I1206 05:43:58.428231 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 05:43:58 crc kubenswrapper[4957]: I1206 05:43:58.647021 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 05:43:58 crc kubenswrapper[4957]: I1206 05:43:58.671117 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 05:43:58 crc kubenswrapper[4957]: I1206 05:43:58.946037 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 05:43:59 crc kubenswrapper[4957]: I1206 05:43:59.188268 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 05:43:59 crc kubenswrapper[4957]: I1206 05:43:59.409924 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 05:43:59 crc kubenswrapper[4957]: I1206 05:43:59.459792 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 05:43:59 crc kubenswrapper[4957]: I1206 05:43:59.581390 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:43:59.595762 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:43:59.763435 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:43:59.884865 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:43:59.949545 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.025801 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.134749 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.187469 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.252251 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.266696 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.315255 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.475210 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.477618 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.574714 4957 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.578112 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=34.578086975 podStartE2EDuration="34.578086975s" podCreationTimestamp="2025-12-06 05:43:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:43:47.818368968 +0000 UTC m=+267.468636610" watchObservedRunningTime="2025-12-06 05:44:00.578086975 +0000 UTC m=+280.228354637" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.583223 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.583320 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.588093 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.612043 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=12.612025918 podStartE2EDuration="12.612025918s" podCreationTimestamp="2025-12-06 05:43:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:44:00.608969812 +0000 UTC m=+280.259237494" watchObservedRunningTime="2025-12-06 05:44:00.612025918 +0000 UTC m=+280.262293550" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.710910 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.711814 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.758479 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.813256 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:00.886882 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.002478 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.012437 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.034720 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.036249 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.082851 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.087453 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.184243 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.319471 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.345093 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.384803 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.421318 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.483555 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.497503 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.509478 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.611972 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:01.612020 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.041172 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.041969 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.043574 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.122496 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.214548 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.255457 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.290123 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.325059 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.353346 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.467316 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.489126 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.497956 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.607938 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.643422 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.800651 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 05:44:02 crc kubenswrapper[4957]: I1206 05:44:02.992632 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.189280 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.351652 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.360502 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.427235 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.517341 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.577453 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.593617 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.636397 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.646398 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.670031 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.708508 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.709794 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.736250 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.759355 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.813363 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.843756 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.866779 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.900603 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.909467 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 05:44:03 crc kubenswrapper[4957]: I1206 05:44:03.969987 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.003119 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.047891 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.163422 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.163681 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.206813 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.289821 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.385203 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.411035 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.438613 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.461190 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.532588 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.730369 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.763091 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.764644 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.778178 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 05:44:04 crc kubenswrapper[4957]: I1206 05:44:04.953147 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.077874 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.320557 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.331821 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.352670 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.382118 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.403797 4957 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.444049 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.491406 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.501952 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.551652 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.622961 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.636463 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.668993 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.805640 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.837584 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.869698 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.882403 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.945094 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.948858 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 05:44:05 crc kubenswrapper[4957]: I1206 05:44:05.981590 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.000080 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.046209 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.087781 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.156458 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.163996 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.194941 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.220239 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.263113 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.331027 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.334378 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.385563 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.385749 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.411476 4957 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.482541 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.498248 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.529006 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.578277 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.611256 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.650209 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.720489 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.727354 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.785095 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.789610 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 05:44:06 crc kubenswrapper[4957]: I1206 05:44:06.793873 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.031188 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.061511 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.143370 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.203341 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.307681 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.323854 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.388077 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.392603 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.538680 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.552729 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.626666 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.672816 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.759706 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.802631 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 05:44:07 crc kubenswrapper[4957]: I1206 05:44:07.996913 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.197542 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.202695 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.250405 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.269588 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.291855 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.335415 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.383435 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.438253 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.506017 4957 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.571517 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.607459 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.618816 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.686265 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.704789 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.746906 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.754776 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.778163 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.790866 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.836188 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.886196 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.948542 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 05:44:08 crc kubenswrapper[4957]: I1206 05:44:08.974488 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.040440 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.055103 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.077490 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.111151 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.116487 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.126993 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.130955 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.155406 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.185933 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.285278 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.357149 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.526547 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.550258 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.560622 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.587671 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.658608 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.772268 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.875494 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 05:44:09 crc kubenswrapper[4957]: I1206 05:44:09.943315 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.012310 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.149408 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.152671 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.244681 4957 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.281047 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.297634 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.368194 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.374554 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.486803 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.526154 4957 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.526458 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e" gracePeriod=5 Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.538960 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.751332 4957 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.802027 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.820549 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.845721 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.872977 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.892410 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 05:44:10 crc kubenswrapper[4957]: I1206 05:44:10.963325 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.007992 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.021821 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.074787 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.178101 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.225112 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.255232 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.406188 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.410776 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.666057 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.841894 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.850302 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 05:44:11 crc kubenswrapper[4957]: I1206 05:44:11.913795 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.081345 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.104674 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.197192 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.441472 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.474262 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.652064 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.705769 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.795466 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 05:44:12 crc kubenswrapper[4957]: I1206 05:44:12.795784 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 05:44:13 crc kubenswrapper[4957]: I1206 05:44:13.160495 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 05:44:13 crc kubenswrapper[4957]: I1206 05:44:13.472826 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 05:44:13 crc kubenswrapper[4957]: I1206 05:44:13.500886 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 05:44:13 crc kubenswrapper[4957]: I1206 05:44:13.804636 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 05:44:14 crc kubenswrapper[4957]: I1206 05:44:14.078786 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 05:44:14 crc kubenswrapper[4957]: I1206 05:44:14.103561 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 05:44:14 crc kubenswrapper[4957]: I1206 05:44:14.146266 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 05:44:14 crc kubenswrapper[4957]: I1206 05:44:14.742910 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 05:44:14 crc kubenswrapper[4957]: I1206 05:44:14.844516 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 05:44:14 crc kubenswrapper[4957]: I1206 05:44:14.851478 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 05:44:15 crc kubenswrapper[4957]: I1206 05:44:15.020027 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 05:44:15 crc kubenswrapper[4957]: I1206 05:44:15.085149 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 05:44:15 crc kubenswrapper[4957]: I1206 05:44:15.307052 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 05:44:15 crc kubenswrapper[4957]: I1206 05:44:15.540757 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.134582 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.135175 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.323752 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.323916 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.323934 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.323970 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.323986 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.324025 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.324125 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.324449 4957 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.324464 4957 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.324508 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.324536 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.335062 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.387323 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.387411 4957 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e" exitCode=137 Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.387491 4957 scope.go:117] "RemoveContainer" containerID="d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.387717 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.426040 4957 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.426123 4957 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.426139 4957 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.426295 4957 scope.go:117] "RemoveContainer" containerID="d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e" Dec 06 05:44:16 crc kubenswrapper[4957]: E1206 05:44:16.426945 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e\": container with ID starting with d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e not found: ID does not exist" containerID="d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.426984 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e"} err="failed to get container status \"d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e\": rpc error: code = NotFound desc = could not find container \"d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e\": container with ID starting with d50b00cf6ec9695719e60e7cefa8f39e529eadae8c474b142a2195478e8fbb3e not found: ID does not exist" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.669520 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.669797 4957 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.681057 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.681106 4957 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="ad80adf3-1f2b-4842-a343-65eedc153a46" Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.685331 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 05:44:16 crc kubenswrapper[4957]: I1206 05:44:16.685578 4957 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="ad80adf3-1f2b-4842-a343-65eedc153a46" Dec 06 05:44:37 crc kubenswrapper[4957]: I1206 05:44:37.342193 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z"] Dec 06 05:44:37 crc kubenswrapper[4957]: I1206 05:44:37.343075 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" podUID="b9e31a64-6ce7-4b24-9891-9775e7193b8a" containerName="route-controller-manager" containerID="cri-o://c721173487db5161059edd627c258646e356fe2a7ca3a6a795787a921a81bc7d" gracePeriod=30 Dec 06 05:44:37 crc kubenswrapper[4957]: I1206 05:44:37.348133 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8m4pc"] Dec 06 05:44:37 crc kubenswrapper[4957]: I1206 05:44:37.348381 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" podUID="950d1218-8492-4736-8dbb-cfa12d544bae" containerName="controller-manager" containerID="cri-o://bfd867228f5dd65e3fbc69a7480136147f047364dc11e4fa4dde4b962cee4468" gracePeriod=30 Dec 06 05:44:37 crc kubenswrapper[4957]: I1206 05:44:37.446499 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.528674 4957 generic.go:334] "Generic (PLEG): container finished" podID="b9e31a64-6ce7-4b24-9891-9775e7193b8a" containerID="c721173487db5161059edd627c258646e356fe2a7ca3a6a795787a921a81bc7d" exitCode=0 Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.528771 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" event={"ID":"b9e31a64-6ce7-4b24-9891-9775e7193b8a","Type":"ContainerDied","Data":"c721173487db5161059edd627c258646e356fe2a7ca3a6a795787a921a81bc7d"} Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.531761 4957 generic.go:334] "Generic (PLEG): container finished" podID="950d1218-8492-4736-8dbb-cfa12d544bae" containerID="bfd867228f5dd65e3fbc69a7480136147f047364dc11e4fa4dde4b962cee4468" exitCode=0 Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.531818 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" event={"ID":"950d1218-8492-4736-8dbb-cfa12d544bae","Type":"ContainerDied","Data":"bfd867228f5dd65e3fbc69a7480136147f047364dc11e4fa4dde4b962cee4468"} Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.531906 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" event={"ID":"950d1218-8492-4736-8dbb-cfa12d544bae","Type":"ContainerDied","Data":"8f27967ed207cb88252514667e6835d4e82f7a678949286146db939edb916f4b"} Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.531925 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f27967ed207cb88252514667e6835d4e82f7a678949286146db939edb916f4b" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.553070 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.573673 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.582031 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.601628 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6dc9b44d9-f6p26"] Dec 06 05:44:38 crc kubenswrapper[4957]: E1206 05:44:38.601860 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e31a64-6ce7-4b24-9891-9775e7193b8a" containerName="route-controller-manager" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.601872 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e31a64-6ce7-4b24-9891-9775e7193b8a" containerName="route-controller-manager" Dec 06 05:44:38 crc kubenswrapper[4957]: E1206 05:44:38.601881 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="950d1218-8492-4736-8dbb-cfa12d544bae" containerName="controller-manager" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.601888 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="950d1218-8492-4736-8dbb-cfa12d544bae" containerName="controller-manager" Dec 06 05:44:38 crc kubenswrapper[4957]: E1206 05:44:38.601902 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" containerName="installer" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.601908 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" containerName="installer" Dec 06 05:44:38 crc kubenswrapper[4957]: E1206 05:44:38.601915 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.601921 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.602004 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9e31a64-6ce7-4b24-9891-9775e7193b8a" containerName="route-controller-manager" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.602017 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e62592-5ba0-42ee-b9a0-41bfa6173821" containerName="installer" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.602027 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="950d1218-8492-4736-8dbb-cfa12d544bae" containerName="controller-manager" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.602035 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.602371 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.621013 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6dc9b44d9-f6p26"] Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.631746 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-config\") pod \"950d1218-8492-4736-8dbb-cfa12d544bae\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.631793 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-proxy-ca-bundles\") pod \"950d1218-8492-4736-8dbb-cfa12d544bae\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.631933 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950d1218-8492-4736-8dbb-cfa12d544bae-serving-cert\") pod \"950d1218-8492-4736-8dbb-cfa12d544bae\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.631964 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fprz\" (UniqueName: \"kubernetes.io/projected/b9e31a64-6ce7-4b24-9891-9775e7193b8a-kube-api-access-5fprz\") pod \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.631999 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-client-ca\") pod \"950d1218-8492-4736-8dbb-cfa12d544bae\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632031 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-config\") pod \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632090 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-client-ca\") pod \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632111 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e31a64-6ce7-4b24-9891-9775e7193b8a-serving-cert\") pod \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\" (UID: \"b9e31a64-6ce7-4b24-9891-9775e7193b8a\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632134 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vkxr\" (UniqueName: \"kubernetes.io/projected/950d1218-8492-4736-8dbb-cfa12d544bae-kube-api-access-6vkxr\") pod \"950d1218-8492-4736-8dbb-cfa12d544bae\" (UID: \"950d1218-8492-4736-8dbb-cfa12d544bae\") " Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632370 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-config\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632407 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9278a3-7c89-400d-a3aa-9ee3a4060617-serving-cert\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632448 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-client-ca\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632471 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-proxy-ca-bundles\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632491 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn68s\" (UniqueName: \"kubernetes.io/projected/ed9278a3-7c89-400d-a3aa-9ee3a4060617-kube-api-access-hn68s\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.632701 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-config" (OuterVolumeSpecName: "config") pod "950d1218-8492-4736-8dbb-cfa12d544bae" (UID: "950d1218-8492-4736-8dbb-cfa12d544bae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.633306 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "950d1218-8492-4736-8dbb-cfa12d544bae" (UID: "950d1218-8492-4736-8dbb-cfa12d544bae"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.634108 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-client-ca" (OuterVolumeSpecName: "client-ca") pod "950d1218-8492-4736-8dbb-cfa12d544bae" (UID: "950d1218-8492-4736-8dbb-cfa12d544bae"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.634392 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-client-ca" (OuterVolumeSpecName: "client-ca") pod "b9e31a64-6ce7-4b24-9891-9775e7193b8a" (UID: "b9e31a64-6ce7-4b24-9891-9775e7193b8a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.634407 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-config" (OuterVolumeSpecName: "config") pod "b9e31a64-6ce7-4b24-9891-9775e7193b8a" (UID: "b9e31a64-6ce7-4b24-9891-9775e7193b8a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.644529 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9e31a64-6ce7-4b24-9891-9775e7193b8a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b9e31a64-6ce7-4b24-9891-9775e7193b8a" (UID: "b9e31a64-6ce7-4b24-9891-9775e7193b8a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.644559 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/950d1218-8492-4736-8dbb-cfa12d544bae-kube-api-access-6vkxr" (OuterVolumeSpecName: "kube-api-access-6vkxr") pod "950d1218-8492-4736-8dbb-cfa12d544bae" (UID: "950d1218-8492-4736-8dbb-cfa12d544bae"). InnerVolumeSpecName "kube-api-access-6vkxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.645184 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/950d1218-8492-4736-8dbb-cfa12d544bae-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "950d1218-8492-4736-8dbb-cfa12d544bae" (UID: "950d1218-8492-4736-8dbb-cfa12d544bae"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.645356 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9e31a64-6ce7-4b24-9891-9775e7193b8a-kube-api-access-5fprz" (OuterVolumeSpecName: "kube-api-access-5fprz") pod "b9e31a64-6ce7-4b24-9891-9775e7193b8a" (UID: "b9e31a64-6ce7-4b24-9891-9775e7193b8a"). InnerVolumeSpecName "kube-api-access-5fprz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.733490 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-config\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.733611 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9278a3-7c89-400d-a3aa-9ee3a4060617-serving-cert\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.733687 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-client-ca\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.734879 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn68s\" (UniqueName: \"kubernetes.io/projected/ed9278a3-7c89-400d-a3aa-9ee3a4060617-kube-api-access-hn68s\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.734935 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-proxy-ca-bundles\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735021 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735043 4957 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735059 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950d1218-8492-4736-8dbb-cfa12d544bae-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735070 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fprz\" (UniqueName: \"kubernetes.io/projected/b9e31a64-6ce7-4b24-9891-9775e7193b8a-kube-api-access-5fprz\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735083 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950d1218-8492-4736-8dbb-cfa12d544bae-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735094 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735105 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9e31a64-6ce7-4b24-9891-9775e7193b8a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735116 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9e31a64-6ce7-4b24-9891-9775e7193b8a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735129 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vkxr\" (UniqueName: \"kubernetes.io/projected/950d1218-8492-4736-8dbb-cfa12d544bae-kube-api-access-6vkxr\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.735434 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-client-ca\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.736207 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-config\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.737038 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-proxy-ca-bundles\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.737924 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9278a3-7c89-400d-a3aa-9ee3a4060617-serving-cert\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.754160 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn68s\" (UniqueName: \"kubernetes.io/projected/ed9278a3-7c89-400d-a3aa-9ee3a4060617-kube-api-access-hn68s\") pod \"controller-manager-6dc9b44d9-f6p26\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:38 crc kubenswrapper[4957]: I1206 05:44:38.915577 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.143398 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6dc9b44d9-f6p26"] Dec 06 05:44:39 crc kubenswrapper[4957]: W1206 05:44:39.153099 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded9278a3_7c89_400d_a3aa_9ee3a4060617.slice/crio-f7797eed9a5d3461c3b363c3060fd3a4bb44dd90a1eb922094ab3d61234782d3 WatchSource:0}: Error finding container f7797eed9a5d3461c3b363c3060fd3a4bb44dd90a1eb922094ab3d61234782d3: Status 404 returned error can't find the container with id f7797eed9a5d3461c3b363c3060fd3a4bb44dd90a1eb922094ab3d61234782d3 Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.537805 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.537801 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z" event={"ID":"b9e31a64-6ce7-4b24-9891-9775e7193b8a","Type":"ContainerDied","Data":"9add6f18901d248524103582db4541bc2679114fb80880dc60f374f5b92152df"} Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.538277 4957 scope.go:117] "RemoveContainer" containerID="c721173487db5161059edd627c258646e356fe2a7ca3a6a795787a921a81bc7d" Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.541576 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" event={"ID":"ed9278a3-7c89-400d-a3aa-9ee3a4060617","Type":"ContainerStarted","Data":"709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b"} Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.541653 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" event={"ID":"ed9278a3-7c89-400d-a3aa-9ee3a4060617","Type":"ContainerStarted","Data":"f7797eed9a5d3461c3b363c3060fd3a4bb44dd90a1eb922094ab3d61234782d3"} Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.541607 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8m4pc" Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.578587 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" podStartSLOduration=2.578565498 podStartE2EDuration="2.578565498s" podCreationTimestamp="2025-12-06 05:44:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:44:39.568483942 +0000 UTC m=+319.218751584" watchObservedRunningTime="2025-12-06 05:44:39.578565498 +0000 UTC m=+319.228833130" Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.580744 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8m4pc"] Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.586107 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8m4pc"] Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.593474 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z"] Dec 06 05:44:39 crc kubenswrapper[4957]: I1206 05:44:39.598356 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vpk8z"] Dec 06 05:44:40 crc kubenswrapper[4957]: I1206 05:44:40.553060 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:40 crc kubenswrapper[4957]: I1206 05:44:40.561410 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:40 crc kubenswrapper[4957]: I1206 05:44:40.675739 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="950d1218-8492-4736-8dbb-cfa12d544bae" path="/var/lib/kubelet/pods/950d1218-8492-4736-8dbb-cfa12d544bae/volumes" Dec 06 05:44:40 crc kubenswrapper[4957]: I1206 05:44:40.676997 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9e31a64-6ce7-4b24-9891-9775e7193b8a" path="/var/lib/kubelet/pods/b9e31a64-6ce7-4b24-9891-9775e7193b8a/volumes" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.175273 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6dc9b44d9-f6p26"] Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.235685 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988"] Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.236582 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.239246 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.239483 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.239661 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.239819 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.240001 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.240327 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.270797 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988"] Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.369419 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-config\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.369531 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-client-ca\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.370287 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57rnq\" (UniqueName: \"kubernetes.io/projected/829c1680-6163-4991-b375-b01ca76ff18c-kube-api-access-57rnq\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.370351 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829c1680-6163-4991-b375-b01ca76ff18c-serving-cert\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.371745 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988"] Dec 06 05:44:41 crc kubenswrapper[4957]: E1206 05:44:41.372609 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-57rnq serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" podUID="829c1680-6163-4991-b375-b01ca76ff18c" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.472545 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57rnq\" (UniqueName: \"kubernetes.io/projected/829c1680-6163-4991-b375-b01ca76ff18c-kube-api-access-57rnq\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.472655 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829c1680-6163-4991-b375-b01ca76ff18c-serving-cert\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.472716 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-config\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.472788 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-client-ca\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.474464 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-client-ca\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.474492 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-config\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.480639 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829c1680-6163-4991-b375-b01ca76ff18c-serving-cert\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.501227 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57rnq\" (UniqueName: \"kubernetes.io/projected/829c1680-6163-4991-b375-b01ca76ff18c-kube-api-access-57rnq\") pod \"route-controller-manager-859874c7b5-g5988\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.560193 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.571946 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.681973 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-config\") pod \"829c1680-6163-4991-b375-b01ca76ff18c\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.682088 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-client-ca\") pod \"829c1680-6163-4991-b375-b01ca76ff18c\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.682148 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57rnq\" (UniqueName: \"kubernetes.io/projected/829c1680-6163-4991-b375-b01ca76ff18c-kube-api-access-57rnq\") pod \"829c1680-6163-4991-b375-b01ca76ff18c\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.682242 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829c1680-6163-4991-b375-b01ca76ff18c-serving-cert\") pod \"829c1680-6163-4991-b375-b01ca76ff18c\" (UID: \"829c1680-6163-4991-b375-b01ca76ff18c\") " Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.682600 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-config" (OuterVolumeSpecName: "config") pod "829c1680-6163-4991-b375-b01ca76ff18c" (UID: "829c1680-6163-4991-b375-b01ca76ff18c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.682868 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-client-ca" (OuterVolumeSpecName: "client-ca") pod "829c1680-6163-4991-b375-b01ca76ff18c" (UID: "829c1680-6163-4991-b375-b01ca76ff18c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.685877 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829c1680-6163-4991-b375-b01ca76ff18c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "829c1680-6163-4991-b375-b01ca76ff18c" (UID: "829c1680-6163-4991-b375-b01ca76ff18c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.686999 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829c1680-6163-4991-b375-b01ca76ff18c-kube-api-access-57rnq" (OuterVolumeSpecName: "kube-api-access-57rnq") pod "829c1680-6163-4991-b375-b01ca76ff18c" (UID: "829c1680-6163-4991-b375-b01ca76ff18c"). InnerVolumeSpecName "kube-api-access-57rnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.784192 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57rnq\" (UniqueName: \"kubernetes.io/projected/829c1680-6163-4991-b375-b01ca76ff18c-kube-api-access-57rnq\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.784242 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/829c1680-6163-4991-b375-b01ca76ff18c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.784268 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.784279 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/829c1680-6163-4991-b375-b01ca76ff18c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:41 crc kubenswrapper[4957]: I1206 05:44:41.953520 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.565678 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.565789 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" podUID="ed9278a3-7c89-400d-a3aa-9ee3a4060617" containerName="controller-manager" containerID="cri-o://709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b" gracePeriod=30 Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.607117 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988"] Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.611582 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq"] Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.612347 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.614190 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.614335 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.614560 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.614682 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.614907 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.616168 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.616274 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-g5988"] Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.620469 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq"] Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.667863 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829c1680-6163-4991-b375-b01ca76ff18c" path="/var/lib/kubelet/pods/829c1680-6163-4991-b375-b01ca76ff18c/volumes" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.699583 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-client-ca\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.700284 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5rd5\" (UniqueName: \"kubernetes.io/projected/d474bedc-89d0-4427-a273-2c710860f436-kube-api-access-f5rd5\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.700337 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-config\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.700375 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d474bedc-89d0-4427-a273-2c710860f436-serving-cert\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.801920 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-client-ca\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.802427 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5rd5\" (UniqueName: \"kubernetes.io/projected/d474bedc-89d0-4427-a273-2c710860f436-kube-api-access-f5rd5\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.802477 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-config\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.802498 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d474bedc-89d0-4427-a273-2c710860f436-serving-cert\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.802810 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-client-ca\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.803869 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-config\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.823629 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d474bedc-89d0-4427-a273-2c710860f436-serving-cert\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.839517 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5rd5\" (UniqueName: \"kubernetes.io/projected/d474bedc-89d0-4427-a273-2c710860f436-kube-api-access-f5rd5\") pod \"route-controller-manager-5c8f48f598-8vfzq\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:42 crc kubenswrapper[4957]: I1206 05:44:42.956523 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.031900 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.106002 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-proxy-ca-bundles\") pod \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.106444 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9278a3-7c89-400d-a3aa-9ee3a4060617-serving-cert\") pod \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.106500 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-config\") pod \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.106582 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-client-ca\") pod \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.106629 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn68s\" (UniqueName: \"kubernetes.io/projected/ed9278a3-7c89-400d-a3aa-9ee3a4060617-kube-api-access-hn68s\") pod \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\" (UID: \"ed9278a3-7c89-400d-a3aa-9ee3a4060617\") " Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.107492 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ed9278a3-7c89-400d-a3aa-9ee3a4060617" (UID: "ed9278a3-7c89-400d-a3aa-9ee3a4060617"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.107515 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-client-ca" (OuterVolumeSpecName: "client-ca") pod "ed9278a3-7c89-400d-a3aa-9ee3a4060617" (UID: "ed9278a3-7c89-400d-a3aa-9ee3a4060617"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.107887 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-config" (OuterVolumeSpecName: "config") pod "ed9278a3-7c89-400d-a3aa-9ee3a4060617" (UID: "ed9278a3-7c89-400d-a3aa-9ee3a4060617"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.109846 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed9278a3-7c89-400d-a3aa-9ee3a4060617-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ed9278a3-7c89-400d-a3aa-9ee3a4060617" (UID: "ed9278a3-7c89-400d-a3aa-9ee3a4060617"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.109911 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9278a3-7c89-400d-a3aa-9ee3a4060617-kube-api-access-hn68s" (OuterVolumeSpecName: "kube-api-access-hn68s") pod "ed9278a3-7c89-400d-a3aa-9ee3a4060617" (UID: "ed9278a3-7c89-400d-a3aa-9ee3a4060617"). InnerVolumeSpecName "kube-api-access-hn68s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.208415 4957 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.208469 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed9278a3-7c89-400d-a3aa-9ee3a4060617-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.208489 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.208501 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed9278a3-7c89-400d-a3aa-9ee3a4060617-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.208516 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn68s\" (UniqueName: \"kubernetes.io/projected/ed9278a3-7c89-400d-a3aa-9ee3a4060617-kube-api-access-hn68s\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.212388 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq"] Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.575022 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" event={"ID":"d474bedc-89d0-4427-a273-2c710860f436","Type":"ContainerStarted","Data":"fd7c6973ff45342384f43b6f4676c076cef73e3969008112f479181f0081c735"} Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.575101 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" event={"ID":"d474bedc-89d0-4427-a273-2c710860f436","Type":"ContainerStarted","Data":"1035e0abcc99ac7bcb833692e6b2ddf967a7da3e3bda71a008dcaff7f5dcfddd"} Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.575649 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.577057 4957 patch_prober.go:28] interesting pod/route-controller-manager-5c8f48f598-8vfzq container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.577108 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" podUID="d474bedc-89d0-4427-a273-2c710860f436" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.577436 4957 generic.go:334] "Generic (PLEG): container finished" podID="ed9278a3-7c89-400d-a3aa-9ee3a4060617" containerID="709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b" exitCode=0 Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.577481 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" event={"ID":"ed9278a3-7c89-400d-a3aa-9ee3a4060617","Type":"ContainerDied","Data":"709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b"} Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.577509 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" event={"ID":"ed9278a3-7c89-400d-a3aa-9ee3a4060617","Type":"ContainerDied","Data":"f7797eed9a5d3461c3b363c3060fd3a4bb44dd90a1eb922094ab3d61234782d3"} Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.577487 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dc9b44d9-f6p26" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.577533 4957 scope.go:117] "RemoveContainer" containerID="709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.599134 4957 scope.go:117] "RemoveContainer" containerID="709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b" Dec 06 05:44:43 crc kubenswrapper[4957]: E1206 05:44:43.601702 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b\": container with ID starting with 709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b not found: ID does not exist" containerID="709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.601767 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b"} err="failed to get container status \"709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b\": rpc error: code = NotFound desc = could not find container \"709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b\": container with ID starting with 709bd81b7f5997c667a5b1f44075e26eb4d72a55b84227d526c325ed676c340b not found: ID does not exist" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.602512 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" podStartSLOduration=2.60248799 podStartE2EDuration="2.60248799s" podCreationTimestamp="2025-12-06 05:44:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:44:43.6000212 +0000 UTC m=+323.250288852" watchObservedRunningTime="2025-12-06 05:44:43.60248799 +0000 UTC m=+323.252755632" Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.624869 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6dc9b44d9-f6p26"] Dec 06 05:44:43 crc kubenswrapper[4957]: I1206 05:44:43.637022 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6dc9b44d9-f6p26"] Dec 06 05:44:44 crc kubenswrapper[4957]: I1206 05:44:44.598002 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:44 crc kubenswrapper[4957]: I1206 05:44:44.676773 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed9278a3-7c89-400d-a3aa-9ee3a4060617" path="/var/lib/kubelet/pods/ed9278a3-7c89-400d-a3aa-9ee3a4060617/volumes" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.293643 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b9d4449c6-kj756"] Dec 06 05:44:45 crc kubenswrapper[4957]: E1206 05:44:45.293905 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9278a3-7c89-400d-a3aa-9ee3a4060617" containerName="controller-manager" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.293925 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9278a3-7c89-400d-a3aa-9ee3a4060617" containerName="controller-manager" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.294094 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9278a3-7c89-400d-a3aa-9ee3a4060617" containerName="controller-manager" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.294553 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.298827 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.299325 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.299707 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.301477 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.302621 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.303154 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.310227 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.310354 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b9d4449c6-kj756"] Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.351243 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-config\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.351670 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-client-ca\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.351708 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28k4q\" (UniqueName: \"kubernetes.io/projected/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-kube-api-access-28k4q\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.351733 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-serving-cert\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.351775 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-proxy-ca-bundles\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.452894 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-config\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.452945 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-client-ca\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.452983 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28k4q\" (UniqueName: \"kubernetes.io/projected/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-kube-api-access-28k4q\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.453012 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-serving-cert\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.453059 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-proxy-ca-bundles\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.454095 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-client-ca\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.454182 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-proxy-ca-bundles\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.455503 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-config\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.462490 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-serving-cert\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.468936 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28k4q\" (UniqueName: \"kubernetes.io/projected/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-kube-api-access-28k4q\") pod \"controller-manager-5b9d4449c6-kj756\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.628329 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:45 crc kubenswrapper[4957]: I1206 05:44:45.824966 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b9d4449c6-kj756"] Dec 06 05:44:45 crc kubenswrapper[4957]: W1206 05:44:45.834583 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b7c9cc2_54e9_40f0_93d5_5513e2c198df.slice/crio-e5fc0e972e1722b00f7cc298e6e85df18fd97c6a1549cc65f96d9cfea4bd4f3a WatchSource:0}: Error finding container e5fc0e972e1722b00f7cc298e6e85df18fd97c6a1549cc65f96d9cfea4bd4f3a: Status 404 returned error can't find the container with id e5fc0e972e1722b00f7cc298e6e85df18fd97c6a1549cc65f96d9cfea4bd4f3a Dec 06 05:44:46 crc kubenswrapper[4957]: I1206 05:44:46.601596 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" event={"ID":"2b7c9cc2-54e9-40f0-93d5-5513e2c198df","Type":"ContainerStarted","Data":"3b5e70b1c01f829cc7b4248d7151d112aae5ce516b2fa12388d8233c24a4f84a"} Dec 06 05:44:46 crc kubenswrapper[4957]: I1206 05:44:46.601672 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" event={"ID":"2b7c9cc2-54e9-40f0-93d5-5513e2c198df","Type":"ContainerStarted","Data":"e5fc0e972e1722b00f7cc298e6e85df18fd97c6a1549cc65f96d9cfea4bd4f3a"} Dec 06 05:44:46 crc kubenswrapper[4957]: I1206 05:44:46.601895 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:46 crc kubenswrapper[4957]: I1206 05:44:46.605880 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:46 crc kubenswrapper[4957]: I1206 05:44:46.624893 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" podStartSLOduration=5.624871569 podStartE2EDuration="5.624871569s" podCreationTimestamp="2025-12-06 05:44:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:44:46.620415134 +0000 UTC m=+326.270682776" watchObservedRunningTime="2025-12-06 05:44:46.624871569 +0000 UTC m=+326.275139221" Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.336095 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b9d4449c6-kj756"] Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.336868 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" podUID="2b7c9cc2-54e9-40f0-93d5-5513e2c198df" containerName="controller-manager" containerID="cri-o://3b5e70b1c01f829cc7b4248d7151d112aae5ce516b2fa12388d8233c24a4f84a" gracePeriod=30 Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.352352 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq"] Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.352576 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" podUID="d474bedc-89d0-4427-a273-2c710860f436" containerName="route-controller-manager" containerID="cri-o://fd7c6973ff45342384f43b6f4676c076cef73e3969008112f479181f0081c735" gracePeriod=30 Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.668636 4957 generic.go:334] "Generic (PLEG): container finished" podID="2b7c9cc2-54e9-40f0-93d5-5513e2c198df" containerID="3b5e70b1c01f829cc7b4248d7151d112aae5ce516b2fa12388d8233c24a4f84a" exitCode=0 Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.668745 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" event={"ID":"2b7c9cc2-54e9-40f0-93d5-5513e2c198df","Type":"ContainerDied","Data":"3b5e70b1c01f829cc7b4248d7151d112aae5ce516b2fa12388d8233c24a4f84a"} Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.671209 4957 generic.go:334] "Generic (PLEG): container finished" podID="d474bedc-89d0-4427-a273-2c710860f436" containerID="fd7c6973ff45342384f43b6f4676c076cef73e3969008112f479181f0081c735" exitCode=0 Dec 06 05:44:57 crc kubenswrapper[4957]: I1206 05:44:57.671251 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" event={"ID":"d474bedc-89d0-4427-a273-2c710860f436","Type":"ContainerDied","Data":"fd7c6973ff45342384f43b6f4676c076cef73e3969008112f479181f0081c735"} Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.383677 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.413170 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2"] Dec 06 05:44:58 crc kubenswrapper[4957]: E1206 05:44:58.413549 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d474bedc-89d0-4427-a273-2c710860f436" containerName="route-controller-manager" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.413574 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="d474bedc-89d0-4427-a273-2c710860f436" containerName="route-controller-manager" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.413690 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="d474bedc-89d0-4427-a273-2c710860f436" containerName="route-controller-manager" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.414315 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.434392 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2"] Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.434554 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-config\") pod \"d474bedc-89d0-4427-a273-2c710860f436\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.434663 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5rd5\" (UniqueName: \"kubernetes.io/projected/d474bedc-89d0-4427-a273-2c710860f436-kube-api-access-f5rd5\") pod \"d474bedc-89d0-4427-a273-2c710860f436\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.434757 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-client-ca\") pod \"d474bedc-89d0-4427-a273-2c710860f436\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.434804 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d474bedc-89d0-4427-a273-2c710860f436-serving-cert\") pod \"d474bedc-89d0-4427-a273-2c710860f436\" (UID: \"d474bedc-89d0-4427-a273-2c710860f436\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.436875 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-client-ca" (OuterVolumeSpecName: "client-ca") pod "d474bedc-89d0-4427-a273-2c710860f436" (UID: "d474bedc-89d0-4427-a273-2c710860f436"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.437345 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-config" (OuterVolumeSpecName: "config") pod "d474bedc-89d0-4427-a273-2c710860f436" (UID: "d474bedc-89d0-4427-a273-2c710860f436"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.443429 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d474bedc-89d0-4427-a273-2c710860f436-kube-api-access-f5rd5" (OuterVolumeSpecName: "kube-api-access-f5rd5") pod "d474bedc-89d0-4427-a273-2c710860f436" (UID: "d474bedc-89d0-4427-a273-2c710860f436"). InnerVolumeSpecName "kube-api-access-f5rd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.451228 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d474bedc-89d0-4427-a273-2c710860f436-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d474bedc-89d0-4427-a273-2c710860f436" (UID: "d474bedc-89d0-4427-a273-2c710860f436"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.536592 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a705dda-3089-40ec-b4ef-12d5a71f1f12-config\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.536653 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xbm2\" (UniqueName: \"kubernetes.io/projected/2a705dda-3089-40ec-b4ef-12d5a71f1f12-kube-api-access-6xbm2\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.536765 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a705dda-3089-40ec-b4ef-12d5a71f1f12-client-ca\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.536916 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a705dda-3089-40ec-b4ef-12d5a71f1f12-serving-cert\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.537000 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.537016 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5rd5\" (UniqueName: \"kubernetes.io/projected/d474bedc-89d0-4427-a273-2c710860f436-kube-api-access-f5rd5\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.537055 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d474bedc-89d0-4427-a273-2c710860f436-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.537076 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d474bedc-89d0-4427-a273-2c710860f436-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.543171 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.637447 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28k4q\" (UniqueName: \"kubernetes.io/projected/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-kube-api-access-28k4q\") pod \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.637585 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-client-ca\") pod \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.637654 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-proxy-ca-bundles\") pod \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.637688 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-serving-cert\") pod \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.637981 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-config\") pod \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\" (UID: \"2b7c9cc2-54e9-40f0-93d5-5513e2c198df\") " Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.638155 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a705dda-3089-40ec-b4ef-12d5a71f1f12-serving-cert\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.638235 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a705dda-3089-40ec-b4ef-12d5a71f1f12-config\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.638263 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xbm2\" (UniqueName: \"kubernetes.io/projected/2a705dda-3089-40ec-b4ef-12d5a71f1f12-kube-api-access-6xbm2\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.638310 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a705dda-3089-40ec-b4ef-12d5a71f1f12-client-ca\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.638713 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-client-ca" (OuterVolumeSpecName: "client-ca") pod "2b7c9cc2-54e9-40f0-93d5-5513e2c198df" (UID: "2b7c9cc2-54e9-40f0-93d5-5513e2c198df"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.638859 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2b7c9cc2-54e9-40f0-93d5-5513e2c198df" (UID: "2b7c9cc2-54e9-40f0-93d5-5513e2c198df"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.638923 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-config" (OuterVolumeSpecName: "config") pod "2b7c9cc2-54e9-40f0-93d5-5513e2c198df" (UID: "2b7c9cc2-54e9-40f0-93d5-5513e2c198df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.639603 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a705dda-3089-40ec-b4ef-12d5a71f1f12-config\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.639750 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a705dda-3089-40ec-b4ef-12d5a71f1f12-client-ca\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.640687 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-kube-api-access-28k4q" (OuterVolumeSpecName: "kube-api-access-28k4q") pod "2b7c9cc2-54e9-40f0-93d5-5513e2c198df" (UID: "2b7c9cc2-54e9-40f0-93d5-5513e2c198df"). InnerVolumeSpecName "kube-api-access-28k4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.643305 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a705dda-3089-40ec-b4ef-12d5a71f1f12-serving-cert\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.645225 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2b7c9cc2-54e9-40f0-93d5-5513e2c198df" (UID: "2b7c9cc2-54e9-40f0-93d5-5513e2c198df"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.654845 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xbm2\" (UniqueName: \"kubernetes.io/projected/2a705dda-3089-40ec-b4ef-12d5a71f1f12-kube-api-access-6xbm2\") pod \"route-controller-manager-859874c7b5-2xlz2\" (UID: \"2a705dda-3089-40ec-b4ef-12d5a71f1f12\") " pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.678002 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" event={"ID":"2b7c9cc2-54e9-40f0-93d5-5513e2c198df","Type":"ContainerDied","Data":"e5fc0e972e1722b00f7cc298e6e85df18fd97c6a1549cc65f96d9cfea4bd4f3a"} Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.678069 4957 scope.go:117] "RemoveContainer" containerID="3b5e70b1c01f829cc7b4248d7151d112aae5ce516b2fa12388d8233c24a4f84a" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.678013 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b9d4449c6-kj756" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.681481 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" event={"ID":"d474bedc-89d0-4427-a273-2c710860f436","Type":"ContainerDied","Data":"1035e0abcc99ac7bcb833692e6b2ddf967a7da3e3bda71a008dcaff7f5dcfddd"} Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.681581 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.693767 4957 scope.go:117] "RemoveContainer" containerID="fd7c6973ff45342384f43b6f4676c076cef73e3969008112f479181f0081c735" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.713613 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b9d4449c6-kj756"] Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.718885 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5b9d4449c6-kj756"] Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.723566 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq"] Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.727936 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8f48f598-8vfzq"] Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.735387 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.740053 4957 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.740096 4957 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.740111 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.740123 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28k4q\" (UniqueName: \"kubernetes.io/projected/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-kube-api-access-28k4q\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.740136 4957 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b7c9cc2-54e9-40f0-93d5-5513e2c198df-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:44:58 crc kubenswrapper[4957]: I1206 05:44:58.920949 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2"] Dec 06 05:44:58 crc kubenswrapper[4957]: W1206 05:44:58.929652 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a705dda_3089_40ec_b4ef_12d5a71f1f12.slice/crio-2da8de3ee6efda44a09c65af3522383536558191ebe2dd39c35542a283ec402e WatchSource:0}: Error finding container 2da8de3ee6efda44a09c65af3522383536558191ebe2dd39c35542a283ec402e: Status 404 returned error can't find the container with id 2da8de3ee6efda44a09c65af3522383536558191ebe2dd39c35542a283ec402e Dec 06 05:44:59 crc kubenswrapper[4957]: I1206 05:44:59.692012 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" event={"ID":"2a705dda-3089-40ec-b4ef-12d5a71f1f12","Type":"ContainerStarted","Data":"6d710aee0cdbb5dbaae5bd7df088ee8cb473809105eb63bec1a3990bbe4d2e82"} Dec 06 05:44:59 crc kubenswrapper[4957]: I1206 05:44:59.692390 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" event={"ID":"2a705dda-3089-40ec-b4ef-12d5a71f1f12","Type":"ContainerStarted","Data":"2da8de3ee6efda44a09c65af3522383536558191ebe2dd39c35542a283ec402e"} Dec 06 05:44:59 crc kubenswrapper[4957]: I1206 05:44:59.692414 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:59 crc kubenswrapper[4957]: I1206 05:44:59.704678 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" Dec 06 05:44:59 crc kubenswrapper[4957]: I1206 05:44:59.715359 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-859874c7b5-2xlz2" podStartSLOduration=2.7153330909999998 podStartE2EDuration="2.715333091s" podCreationTimestamp="2025-12-06 05:44:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:44:59.710239148 +0000 UTC m=+339.360506790" watchObservedRunningTime="2025-12-06 05:44:59.715333091 +0000 UTC m=+339.365600743" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.164844 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4"] Dec 06 05:45:00 crc kubenswrapper[4957]: E1206 05:45:00.165094 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7c9cc2-54e9-40f0-93d5-5513e2c198df" containerName="controller-manager" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.165107 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7c9cc2-54e9-40f0-93d5-5513e2c198df" containerName="controller-manager" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.165207 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7c9cc2-54e9-40f0-93d5-5513e2c198df" containerName="controller-manager" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.165596 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.168252 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.168777 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.174725 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4"] Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.269132 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572c71f1-84ce-4b36-bd03-e17ee311d789-secret-volume\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.269205 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572c71f1-84ce-4b36-bd03-e17ee311d789-config-volume\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.269269 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sphc\" (UniqueName: \"kubernetes.io/projected/572c71f1-84ce-4b36-bd03-e17ee311d789-kube-api-access-4sphc\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.370247 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sphc\" (UniqueName: \"kubernetes.io/projected/572c71f1-84ce-4b36-bd03-e17ee311d789-kube-api-access-4sphc\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.370323 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572c71f1-84ce-4b36-bd03-e17ee311d789-secret-volume\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.370378 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572c71f1-84ce-4b36-bd03-e17ee311d789-config-volume\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.371655 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572c71f1-84ce-4b36-bd03-e17ee311d789-config-volume\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.379694 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572c71f1-84ce-4b36-bd03-e17ee311d789-secret-volume\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.394034 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sphc\" (UniqueName: \"kubernetes.io/projected/572c71f1-84ce-4b36-bd03-e17ee311d789-kube-api-access-4sphc\") pod \"collect-profiles-29416665-t95k4\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.498759 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.672275 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b7c9cc2-54e9-40f0-93d5-5513e2c198df" path="/var/lib/kubelet/pods/2b7c9cc2-54e9-40f0-93d5-5513e2c198df/volumes" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.673931 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d474bedc-89d0-4427-a273-2c710860f436" path="/var/lib/kubelet/pods/d474bedc-89d0-4427-a273-2c710860f436/volumes" Dec 06 05:45:00 crc kubenswrapper[4957]: I1206 05:45:00.751299 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4"] Dec 06 05:45:00 crc kubenswrapper[4957]: W1206 05:45:00.758234 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod572c71f1_84ce_4b36_bd03_e17ee311d789.slice/crio-3807da4276cd28bc11867e16a9338bcdfceb4c065792570c000a4fd9fc787eed WatchSource:0}: Error finding container 3807da4276cd28bc11867e16a9338bcdfceb4c065792570c000a4fd9fc787eed: Status 404 returned error can't find the container with id 3807da4276cd28bc11867e16a9338bcdfceb4c065792570c000a4fd9fc787eed Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.303101 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d57c4f687-hgxvk"] Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.303953 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.308288 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.308810 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.310951 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.311026 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.312583 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.312705 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.316508 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.318694 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d57c4f687-hgxvk"] Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.381669 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-client-ca\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.381716 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r9g4\" (UniqueName: \"kubernetes.io/projected/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-kube-api-access-6r9g4\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.381743 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-config\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.381779 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-serving-cert\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.381847 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-proxy-ca-bundles\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.483657 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-client-ca\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.483712 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r9g4\" (UniqueName: \"kubernetes.io/projected/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-kube-api-access-6r9g4\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.483738 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-config\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.483779 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-serving-cert\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.483813 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-proxy-ca-bundles\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.485178 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-proxy-ca-bundles\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.485185 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-client-ca\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.485493 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-config\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.498645 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-serving-cert\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.503064 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r9g4\" (UniqueName: \"kubernetes.io/projected/25ff1c3c-e565-4ce9-b023-72636d9ed1a0-kube-api-access-6r9g4\") pod \"controller-manager-5d57c4f687-hgxvk\" (UID: \"25ff1c3c-e565-4ce9-b023-72636d9ed1a0\") " pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.656901 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.708806 4957 generic.go:334] "Generic (PLEG): container finished" podID="572c71f1-84ce-4b36-bd03-e17ee311d789" containerID="0df9fa0db471f32867dc2abed3c36956d65ab2c6238d4d76ddc9b930d74090c4" exitCode=0 Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.709760 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" event={"ID":"572c71f1-84ce-4b36-bd03-e17ee311d789","Type":"ContainerDied","Data":"0df9fa0db471f32867dc2abed3c36956d65ab2c6238d4d76ddc9b930d74090c4"} Dec 06 05:45:01 crc kubenswrapper[4957]: I1206 05:45:01.709793 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" event={"ID":"572c71f1-84ce-4b36-bd03-e17ee311d789","Type":"ContainerStarted","Data":"3807da4276cd28bc11867e16a9338bcdfceb4c065792570c000a4fd9fc787eed"} Dec 06 05:45:02 crc kubenswrapper[4957]: I1206 05:45:02.064617 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d57c4f687-hgxvk"] Dec 06 05:45:02 crc kubenswrapper[4957]: I1206 05:45:02.715537 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" event={"ID":"25ff1c3c-e565-4ce9-b023-72636d9ed1a0","Type":"ContainerStarted","Data":"3b2249f40a10c3386305c73cc152d6928ba30edc3cf6b7334fa8b2d381a1bb44"} Dec 06 05:45:02 crc kubenswrapper[4957]: I1206 05:45:02.716003 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" event={"ID":"25ff1c3c-e565-4ce9-b023-72636d9ed1a0","Type":"ContainerStarted","Data":"542acc6655ca01bc204760e936f1a2909674a0cdf827d8dafde60bccb222f6e7"} Dec 06 05:45:02 crc kubenswrapper[4957]: I1206 05:45:02.754231 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" podStartSLOduration=5.754203668 podStartE2EDuration="5.754203668s" podCreationTimestamp="2025-12-06 05:44:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:45:02.75249216 +0000 UTC m=+342.402759802" watchObservedRunningTime="2025-12-06 05:45:02.754203668 +0000 UTC m=+342.404471300" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.056960 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.107023 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sphc\" (UniqueName: \"kubernetes.io/projected/572c71f1-84ce-4b36-bd03-e17ee311d789-kube-api-access-4sphc\") pod \"572c71f1-84ce-4b36-bd03-e17ee311d789\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.107095 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572c71f1-84ce-4b36-bd03-e17ee311d789-secret-volume\") pod \"572c71f1-84ce-4b36-bd03-e17ee311d789\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.107124 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572c71f1-84ce-4b36-bd03-e17ee311d789-config-volume\") pod \"572c71f1-84ce-4b36-bd03-e17ee311d789\" (UID: \"572c71f1-84ce-4b36-bd03-e17ee311d789\") " Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.108324 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/572c71f1-84ce-4b36-bd03-e17ee311d789-config-volume" (OuterVolumeSpecName: "config-volume") pod "572c71f1-84ce-4b36-bd03-e17ee311d789" (UID: "572c71f1-84ce-4b36-bd03-e17ee311d789"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.113321 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572c71f1-84ce-4b36-bd03-e17ee311d789-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "572c71f1-84ce-4b36-bd03-e17ee311d789" (UID: "572c71f1-84ce-4b36-bd03-e17ee311d789"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.113306 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572c71f1-84ce-4b36-bd03-e17ee311d789-kube-api-access-4sphc" (OuterVolumeSpecName: "kube-api-access-4sphc") pod "572c71f1-84ce-4b36-bd03-e17ee311d789" (UID: "572c71f1-84ce-4b36-bd03-e17ee311d789"). InnerVolumeSpecName "kube-api-access-4sphc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.209304 4957 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572c71f1-84ce-4b36-bd03-e17ee311d789-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.209363 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572c71f1-84ce-4b36-bd03-e17ee311d789-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.209377 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sphc\" (UniqueName: \"kubernetes.io/projected/572c71f1-84ce-4b36-bd03-e17ee311d789-kube-api-access-4sphc\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.729096 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" event={"ID":"572c71f1-84ce-4b36-bd03-e17ee311d789","Type":"ContainerDied","Data":"3807da4276cd28bc11867e16a9338bcdfceb4c065792570c000a4fd9fc787eed"} Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.729666 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3807da4276cd28bc11867e16a9338bcdfceb4c065792570c000a4fd9fc787eed" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.729731 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.729812 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4" Dec 06 05:45:03 crc kubenswrapper[4957]: I1206 05:45:03.734729 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d57c4f687-hgxvk" Dec 06 05:45:10 crc kubenswrapper[4957]: I1206 05:45:10.180904 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:45:10 crc kubenswrapper[4957]: I1206 05:45:10.181261 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:45:11 crc kubenswrapper[4957]: I1206 05:45:11.347045 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hz4nt"] Dec 06 05:45:11 crc kubenswrapper[4957]: I1206 05:45:11.347816 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hz4nt" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="registry-server" containerID="cri-o://31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762" gracePeriod=2 Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.760320 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.798378 4957 generic.go:334] "Generic (PLEG): container finished" podID="b5b5df82-a892-4971-84e3-3c3690960f71" containerID="31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762" exitCode=0 Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.798465 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hz4nt" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.798470 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hz4nt" event={"ID":"b5b5df82-a892-4971-84e3-3c3690960f71","Type":"ContainerDied","Data":"31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762"} Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.798569 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hz4nt" event={"ID":"b5b5df82-a892-4971-84e3-3c3690960f71","Type":"ContainerDied","Data":"5cbff16ccfe7539ba8ca31e6ad2be89bd640862e8460ade6beb112919c7b2e70"} Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.798598 4957 scope.go:117] "RemoveContainer" containerID="31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.825463 4957 scope.go:117] "RemoveContainer" containerID="e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.847154 4957 scope.go:117] "RemoveContainer" containerID="b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.858199 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjgb6\" (UniqueName: \"kubernetes.io/projected/b5b5df82-a892-4971-84e3-3c3690960f71-kube-api-access-pjgb6\") pod \"b5b5df82-a892-4971-84e3-3c3690960f71\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.858374 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-utilities\") pod \"b5b5df82-a892-4971-84e3-3c3690960f71\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.858538 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-catalog-content\") pod \"b5b5df82-a892-4971-84e3-3c3690960f71\" (UID: \"b5b5df82-a892-4971-84e3-3c3690960f71\") " Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.859565 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-utilities" (OuterVolumeSpecName: "utilities") pod "b5b5df82-a892-4971-84e3-3c3690960f71" (UID: "b5b5df82-a892-4971-84e3-3c3690960f71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.864252 4957 scope.go:117] "RemoveContainer" containerID="31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762" Dec 06 05:45:13 crc kubenswrapper[4957]: E1206 05:45:13.864776 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762\": container with ID starting with 31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762 not found: ID does not exist" containerID="31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.864861 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762"} err="failed to get container status \"31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762\": rpc error: code = NotFound desc = could not find container \"31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762\": container with ID starting with 31e0af7dad57fbdee02627bc5c49a3c975a596351795ff254c10cce324517762 not found: ID does not exist" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.864913 4957 scope.go:117] "RemoveContainer" containerID="e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8" Dec 06 05:45:13 crc kubenswrapper[4957]: E1206 05:45:13.865378 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8\": container with ID starting with e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8 not found: ID does not exist" containerID="e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.865427 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8"} err="failed to get container status \"e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8\": rpc error: code = NotFound desc = could not find container \"e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8\": container with ID starting with e8f4c13207bbbee05c8002cff1c163739220582ba3ad8d9b2a2369c9c2ea9ba8 not found: ID does not exist" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.865461 4957 scope.go:117] "RemoveContainer" containerID="b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab" Dec 06 05:45:13 crc kubenswrapper[4957]: E1206 05:45:13.865789 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab\": container with ID starting with b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab not found: ID does not exist" containerID="b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.865823 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab"} err="failed to get container status \"b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab\": rpc error: code = NotFound desc = could not find container \"b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab\": container with ID starting with b17488369e383167eb2461383ecc90153269040d165a38da0b5860e81e1f19ab not found: ID does not exist" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.867459 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b5df82-a892-4971-84e3-3c3690960f71-kube-api-access-pjgb6" (OuterVolumeSpecName: "kube-api-access-pjgb6") pod "b5b5df82-a892-4971-84e3-3c3690960f71" (UID: "b5b5df82-a892-4971-84e3-3c3690960f71"). InnerVolumeSpecName "kube-api-access-pjgb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.960810 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjgb6\" (UniqueName: \"kubernetes.io/projected/b5b5df82-a892-4971-84e3-3c3690960f71-kube-api-access-pjgb6\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.960894 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:13 crc kubenswrapper[4957]: I1206 05:45:13.996523 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5b5df82-a892-4971-84e3-3c3690960f71" (UID: "b5b5df82-a892-4971-84e3-3c3690960f71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:14 crc kubenswrapper[4957]: I1206 05:45:14.063184 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b5df82-a892-4971-84e3-3c3690960f71-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:14 crc kubenswrapper[4957]: I1206 05:45:14.154378 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hz4nt"] Dec 06 05:45:14 crc kubenswrapper[4957]: I1206 05:45:14.162992 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hz4nt"] Dec 06 05:45:14 crc kubenswrapper[4957]: I1206 05:45:14.675551 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" path="/var/lib/kubelet/pods/b5b5df82-a892-4971-84e3-3c3690960f71/volumes" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.231189 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rzgbv"] Dec 06 05:45:23 crc kubenswrapper[4957]: E1206 05:45:23.232046 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572c71f1-84ce-4b36-bd03-e17ee311d789" containerName="collect-profiles" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.232063 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="572c71f1-84ce-4b36-bd03-e17ee311d789" containerName="collect-profiles" Dec 06 05:45:23 crc kubenswrapper[4957]: E1206 05:45:23.232074 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="extract-utilities" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.232080 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="extract-utilities" Dec 06 05:45:23 crc kubenswrapper[4957]: E1206 05:45:23.232094 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="registry-server" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.232100 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="registry-server" Dec 06 05:45:23 crc kubenswrapper[4957]: E1206 05:45:23.232109 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="extract-content" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.232114 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="extract-content" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.232222 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="572c71f1-84ce-4b36-bd03-e17ee311d789" containerName="collect-profiles" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.232238 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b5df82-a892-4971-84e3-3c3690960f71" containerName="registry-server" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.232658 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.244757 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rzgbv"] Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286170 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7f3106ac-361a-4829-97d7-9b31f8132446-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286223 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-bound-sa-token\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286253 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-registry-tls\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286394 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7f3106ac-361a-4829-97d7-9b31f8132446-registry-certificates\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286532 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw8gf\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-kube-api-access-mw8gf\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286614 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286644 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f3106ac-361a-4829-97d7-9b31f8132446-trusted-ca\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.286704 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7f3106ac-361a-4829-97d7-9b31f8132446-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.309726 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.388561 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7f3106ac-361a-4829-97d7-9b31f8132446-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.388624 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-bound-sa-token\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.388668 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-registry-tls\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.388709 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7f3106ac-361a-4829-97d7-9b31f8132446-registry-certificates\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.388755 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw8gf\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-kube-api-access-mw8gf\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.388789 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f3106ac-361a-4829-97d7-9b31f8132446-trusted-ca\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.388858 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7f3106ac-361a-4829-97d7-9b31f8132446-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.389886 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7f3106ac-361a-4829-97d7-9b31f8132446-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.390610 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f3106ac-361a-4829-97d7-9b31f8132446-trusted-ca\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.390855 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7f3106ac-361a-4829-97d7-9b31f8132446-registry-certificates\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.398701 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7f3106ac-361a-4829-97d7-9b31f8132446-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.401180 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-registry-tls\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.405864 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-bound-sa-token\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.411737 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw8gf\" (UniqueName: \"kubernetes.io/projected/7f3106ac-361a-4829-97d7-9b31f8132446-kube-api-access-mw8gf\") pod \"image-registry-66df7c8f76-rzgbv\" (UID: \"7f3106ac-361a-4829-97d7-9b31f8132446\") " pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.550365 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:23 crc kubenswrapper[4957]: I1206 05:45:23.990705 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rzgbv"] Dec 06 05:45:23 crc kubenswrapper[4957]: W1206 05:45:23.990827 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f3106ac_361a_4829_97d7_9b31f8132446.slice/crio-e66b9a4a41b46d301114e5c5ec3200eedb80efea34d4819fd16eb0d54b54eb68 WatchSource:0}: Error finding container e66b9a4a41b46d301114e5c5ec3200eedb80efea34d4819fd16eb0d54b54eb68: Status 404 returned error can't find the container with id e66b9a4a41b46d301114e5c5ec3200eedb80efea34d4819fd16eb0d54b54eb68 Dec 06 05:45:24 crc kubenswrapper[4957]: I1206 05:45:24.877285 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" event={"ID":"7f3106ac-361a-4829-97d7-9b31f8132446","Type":"ContainerStarted","Data":"fccad53940f666bf291ee9c3a1447ae3f3b118f943920ade047daca871aaeb7a"} Dec 06 05:45:24 crc kubenswrapper[4957]: I1206 05:45:24.877541 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" event={"ID":"7f3106ac-361a-4829-97d7-9b31f8132446","Type":"ContainerStarted","Data":"e66b9a4a41b46d301114e5c5ec3200eedb80efea34d4819fd16eb0d54b54eb68"} Dec 06 05:45:24 crc kubenswrapper[4957]: I1206 05:45:24.877581 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:24 crc kubenswrapper[4957]: I1206 05:45:24.909983 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" podStartSLOduration=1.909962599 podStartE2EDuration="1.909962599s" podCreationTimestamp="2025-12-06 05:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:45:24.903500248 +0000 UTC m=+364.553767920" watchObservedRunningTime="2025-12-06 05:45:24.909962599 +0000 UTC m=+364.560230231" Dec 06 05:45:40 crc kubenswrapper[4957]: I1206 05:45:40.181660 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:45:40 crc kubenswrapper[4957]: I1206 05:45:40.182567 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.039447 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-44rx8"] Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.040076 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-44rx8" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="registry-server" containerID="cri-o://b22629efd03338ca3b44ce4baab025ae3291b9cce02a093fcb29576f25a525c9" gracePeriod=30 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.053905 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbk94"] Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.054153 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kbk94" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="registry-server" containerID="cri-o://93ba862f38abdd563c612d142e371238f4ba4e805eb9cc42e30f51497e9e1181" gracePeriod=30 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.100461 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gfv9x"] Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.100656 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" podUID="665360fe-3cbc-49bb-9316-01a6cf62174e" containerName="marketplace-operator" containerID="cri-o://c71168b6de9b9d702c8a039537e6c5862ef60811a4ea53f8d23db5ef00314d83" gracePeriod=30 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.139912 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2f5bt"] Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.140204 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2f5bt" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="registry-server" containerID="cri-o://654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d" gracePeriod=30 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.167710 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mzt5s"] Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.168117 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mzt5s" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="registry-server" containerID="cri-o://6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf" gracePeriod=30 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.174013 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cbbgh"] Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.174653 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.184903 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cbbgh"] Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.323144 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.323372 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.323411 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvxlg\" (UniqueName: \"kubernetes.io/projected/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-kube-api-access-fvxlg\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.424573 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.424617 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvxlg\" (UniqueName: \"kubernetes.io/projected/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-kube-api-access-fvxlg\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.424685 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.425975 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.435697 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.439865 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvxlg\" (UniqueName: \"kubernetes.io/projected/de45cbbb-4070-4df4-8b0e-c6dbe32a1a33-kube-api-access-fvxlg\") pod \"marketplace-operator-79b997595-cbbgh\" (UID: \"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33\") " pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.615298 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.637410 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.647711 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.731414 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n2sg\" (UniqueName: \"kubernetes.io/projected/37639dbf-8f08-428e-93e2-456e2946333f-kube-api-access-5n2sg\") pod \"37639dbf-8f08-428e-93e2-456e2946333f\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.731493 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-utilities\") pod \"37639dbf-8f08-428e-93e2-456e2946333f\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.731520 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-catalog-content\") pod \"37639dbf-8f08-428e-93e2-456e2946333f\" (UID: \"37639dbf-8f08-428e-93e2-456e2946333f\") " Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.732821 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-utilities" (OuterVolumeSpecName: "utilities") pod "37639dbf-8f08-428e-93e2-456e2946333f" (UID: "37639dbf-8f08-428e-93e2-456e2946333f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.740139 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37639dbf-8f08-428e-93e2-456e2946333f-kube-api-access-5n2sg" (OuterVolumeSpecName: "kube-api-access-5n2sg") pod "37639dbf-8f08-428e-93e2-456e2946333f" (UID: "37639dbf-8f08-428e-93e2-456e2946333f"). InnerVolumeSpecName "kube-api-access-5n2sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.761604 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37639dbf-8f08-428e-93e2-456e2946333f" (UID: "37639dbf-8f08-428e-93e2-456e2946333f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.833223 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-catalog-content\") pod \"f6be4201-af3d-4f37-a819-7cf136f1e906\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.833260 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-utilities\") pod \"f6be4201-af3d-4f37-a819-7cf136f1e906\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.833309 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrggc\" (UniqueName: \"kubernetes.io/projected/f6be4201-af3d-4f37-a819-7cf136f1e906-kube-api-access-mrggc\") pod \"f6be4201-af3d-4f37-a819-7cf136f1e906\" (UID: \"f6be4201-af3d-4f37-a819-7cf136f1e906\") " Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.833563 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n2sg\" (UniqueName: \"kubernetes.io/projected/37639dbf-8f08-428e-93e2-456e2946333f-kube-api-access-5n2sg\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.833579 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.833590 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37639dbf-8f08-428e-93e2-456e2946333f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.840079 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6be4201-af3d-4f37-a819-7cf136f1e906-kube-api-access-mrggc" (OuterVolumeSpecName: "kube-api-access-mrggc") pod "f6be4201-af3d-4f37-a819-7cf136f1e906" (UID: "f6be4201-af3d-4f37-a819-7cf136f1e906"). InnerVolumeSpecName "kube-api-access-mrggc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.844472 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-utilities" (OuterVolumeSpecName: "utilities") pod "f6be4201-af3d-4f37-a819-7cf136f1e906" (UID: "f6be4201-af3d-4f37-a819-7cf136f1e906"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.935211 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrggc\" (UniqueName: \"kubernetes.io/projected/f6be4201-af3d-4f37-a819-7cf136f1e906-kube-api-access-mrggc\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.935253 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.959709 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6be4201-af3d-4f37-a819-7cf136f1e906" (UID: "f6be4201-af3d-4f37-a819-7cf136f1e906"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.975413 4957 generic.go:334] "Generic (PLEG): container finished" podID="37639dbf-8f08-428e-93e2-456e2946333f" containerID="654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d" exitCode=0 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.975504 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2f5bt" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.975930 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2f5bt" event={"ID":"37639dbf-8f08-428e-93e2-456e2946333f","Type":"ContainerDied","Data":"654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d"} Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.976008 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2f5bt" event={"ID":"37639dbf-8f08-428e-93e2-456e2946333f","Type":"ContainerDied","Data":"462b207bd97a6c6d30b332b360e18ac92dbc7fda833b6338f5eec0ff80bd41f3"} Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.976042 4957 scope.go:117] "RemoveContainer" containerID="654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.977505 4957 generic.go:334] "Generic (PLEG): container finished" podID="665360fe-3cbc-49bb-9316-01a6cf62174e" containerID="c71168b6de9b9d702c8a039537e6c5862ef60811a4ea53f8d23db5ef00314d83" exitCode=0 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.977604 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" event={"ID":"665360fe-3cbc-49bb-9316-01a6cf62174e","Type":"ContainerDied","Data":"c71168b6de9b9d702c8a039537e6c5862ef60811a4ea53f8d23db5ef00314d83"} Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.980606 4957 generic.go:334] "Generic (PLEG): container finished" podID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerID="93ba862f38abdd563c612d142e371238f4ba4e805eb9cc42e30f51497e9e1181" exitCode=0 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.980652 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbk94" event={"ID":"5de294a0-fcab-43a3-8001-c32c3965f3f6","Type":"ContainerDied","Data":"93ba862f38abdd563c612d142e371238f4ba4e805eb9cc42e30f51497e9e1181"} Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.984265 4957 generic.go:334] "Generic (PLEG): container finished" podID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerID="6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf" exitCode=0 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.984310 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mzt5s" Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.984359 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzt5s" event={"ID":"f6be4201-af3d-4f37-a819-7cf136f1e906","Type":"ContainerDied","Data":"6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf"} Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.984414 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzt5s" event={"ID":"f6be4201-af3d-4f37-a819-7cf136f1e906","Type":"ContainerDied","Data":"4d12c532ee11e1fea2995921d7685e330b63c5c2a1a7c5f5efd2561f6b2fc714"} Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.986933 4957 generic.go:334] "Generic (PLEG): container finished" podID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerID="b22629efd03338ca3b44ce4baab025ae3291b9cce02a093fcb29576f25a525c9" exitCode=0 Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.986974 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44rx8" event={"ID":"0dc111dd-bc98-4dfc-8130-85c33d2271b5","Type":"ContainerDied","Data":"b22629efd03338ca3b44ce4baab025ae3291b9cce02a093fcb29576f25a525c9"} Dec 06 05:45:41 crc kubenswrapper[4957]: I1206 05:45:41.995397 4957 scope.go:117] "RemoveContainer" containerID="43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.021755 4957 scope.go:117] "RemoveContainer" containerID="3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.027714 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2f5bt"] Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.031998 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2f5bt"] Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.036548 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be4201-af3d-4f37-a819-7cf136f1e906-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.041689 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mzt5s"] Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.045263 4957 scope.go:117] "RemoveContainer" containerID="654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d" Dec 06 05:45:42 crc kubenswrapper[4957]: E1206 05:45:42.046023 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d\": container with ID starting with 654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d not found: ID does not exist" containerID="654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.046067 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d"} err="failed to get container status \"654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d\": rpc error: code = NotFound desc = could not find container \"654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d\": container with ID starting with 654f24e9c41760161515f8b1b22cc1133fe18d3be33ee16dc549413507e84b3d not found: ID does not exist" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.046094 4957 scope.go:117] "RemoveContainer" containerID="43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912" Dec 06 05:45:42 crc kubenswrapper[4957]: E1206 05:45:42.046532 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912\": container with ID starting with 43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912 not found: ID does not exist" containerID="43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.046555 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912"} err="failed to get container status \"43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912\": rpc error: code = NotFound desc = could not find container \"43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912\": container with ID starting with 43ef024489976dadeee0ca5f920d78a4150ce2b1f80ed7e28dc756aad2d87912 not found: ID does not exist" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.046572 4957 scope.go:117] "RemoveContainer" containerID="3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211" Dec 06 05:45:42 crc kubenswrapper[4957]: E1206 05:45:42.046966 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211\": container with ID starting with 3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211 not found: ID does not exist" containerID="3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.047004 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211"} err="failed to get container status \"3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211\": rpc error: code = NotFound desc = could not find container \"3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211\": container with ID starting with 3e6c960df4514afc213119067d50d52017d05264e9a08198c94022d5a550c211 not found: ID does not exist" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.047027 4957 scope.go:117] "RemoveContainer" containerID="6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.051004 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mzt5s"] Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.065460 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cbbgh"] Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.069046 4957 scope.go:117] "RemoveContainer" containerID="2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8" Dec 06 05:45:42 crc kubenswrapper[4957]: W1206 05:45:42.072452 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde45cbbb_4070_4df4_8b0e_c6dbe32a1a33.slice/crio-8a6dc1b4a6b60561de0d6bfc7e83df989435194c52f9a34ddca06f3877c051f0 WatchSource:0}: Error finding container 8a6dc1b4a6b60561de0d6bfc7e83df989435194c52f9a34ddca06f3877c051f0: Status 404 returned error can't find the container with id 8a6dc1b4a6b60561de0d6bfc7e83df989435194c52f9a34ddca06f3877c051f0 Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.090324 4957 scope.go:117] "RemoveContainer" containerID="dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.113668 4957 scope.go:117] "RemoveContainer" containerID="6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf" Dec 06 05:45:42 crc kubenswrapper[4957]: E1206 05:45:42.114145 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf\": container with ID starting with 6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf not found: ID does not exist" containerID="6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.114205 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf"} err="failed to get container status \"6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf\": rpc error: code = NotFound desc = could not find container \"6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf\": container with ID starting with 6ee4aced2bf12ee7c9b1933be453cb817bc503e308e38df7b7e401ed944ef1cf not found: ID does not exist" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.114232 4957 scope.go:117] "RemoveContainer" containerID="2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8" Dec 06 05:45:42 crc kubenswrapper[4957]: E1206 05:45:42.114540 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8\": container with ID starting with 2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8 not found: ID does not exist" containerID="2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.114594 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8"} err="failed to get container status \"2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8\": rpc error: code = NotFound desc = could not find container \"2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8\": container with ID starting with 2e1abcd3fc2bcf77c348e72adfbf6a89c724edabf6748cbe55cdbe4526d470f8 not found: ID does not exist" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.114612 4957 scope.go:117] "RemoveContainer" containerID="dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4" Dec 06 05:45:42 crc kubenswrapper[4957]: E1206 05:45:42.115050 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4\": container with ID starting with dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4 not found: ID does not exist" containerID="dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.115120 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4"} err="failed to get container status \"dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4\": rpc error: code = NotFound desc = could not find container \"dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4\": container with ID starting with dce8685a42dfb24579da58ecdf29a35e4098d0ae6e865352496037bf784ba8a4 not found: ID does not exist" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.669340 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37639dbf-8f08-428e-93e2-456e2946333f" path="/var/lib/kubelet/pods/37639dbf-8f08-428e-93e2-456e2946333f/volumes" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.672079 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" path="/var/lib/kubelet/pods/f6be4201-af3d-4f37-a819-7cf136f1e906/volumes" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.950688 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.995730 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" event={"ID":"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33","Type":"ContainerStarted","Data":"8a6dc1b4a6b60561de0d6bfc7e83df989435194c52f9a34ddca06f3877c051f0"} Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.998129 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbk94" event={"ID":"5de294a0-fcab-43a3-8001-c32c3965f3f6","Type":"ContainerDied","Data":"f8f8db63e6e74dd360753249b7d6db258bbfe99ea610194b80d31db13da8b292"} Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.998173 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbk94" Dec 06 05:45:42 crc kubenswrapper[4957]: I1206 05:45:42.998282 4957 scope.go:117] "RemoveContainer" containerID="93ba862f38abdd563c612d142e371238f4ba4e805eb9cc42e30f51497e9e1181" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.013948 4957 scope.go:117] "RemoveContainer" containerID="fe2de32d3a062005f8e807692ecd618f813d43790719dfecdcd4ee94d72f8e20" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.029689 4957 scope.go:117] "RemoveContainer" containerID="57c33eb9f88adc9309321905995d7c3532105bf50b257194b8ed8005a53d4890" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.050064 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-utilities\") pod \"5de294a0-fcab-43a3-8001-c32c3965f3f6\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.050178 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-catalog-content\") pod \"5de294a0-fcab-43a3-8001-c32c3965f3f6\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.050225 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckm4d\" (UniqueName: \"kubernetes.io/projected/5de294a0-fcab-43a3-8001-c32c3965f3f6-kube-api-access-ckm4d\") pod \"5de294a0-fcab-43a3-8001-c32c3965f3f6\" (UID: \"5de294a0-fcab-43a3-8001-c32c3965f3f6\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.051825 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-utilities" (OuterVolumeSpecName: "utilities") pod "5de294a0-fcab-43a3-8001-c32c3965f3f6" (UID: "5de294a0-fcab-43a3-8001-c32c3965f3f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.056895 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de294a0-fcab-43a3-8001-c32c3965f3f6-kube-api-access-ckm4d" (OuterVolumeSpecName: "kube-api-access-ckm4d") pod "5de294a0-fcab-43a3-8001-c32c3965f3f6" (UID: "5de294a0-fcab-43a3-8001-c32c3965f3f6"). InnerVolumeSpecName "kube-api-access-ckm4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.109562 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5de294a0-fcab-43a3-8001-c32c3965f3f6" (UID: "5de294a0-fcab-43a3-8001-c32c3965f3f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.152224 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.152320 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5de294a0-fcab-43a3-8001-c32c3965f3f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.152393 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckm4d\" (UniqueName: \"kubernetes.io/projected/5de294a0-fcab-43a3-8001-c32c3965f3f6-kube-api-access-ckm4d\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.377547 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbk94"] Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.381711 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kbk94"] Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.555484 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-rzgbv" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.580003 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.629849 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xlz55"] Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.635369 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.659884 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-catalog-content\") pod \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.659963 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-utilities\") pod \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.660026 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brs4f\" (UniqueName: \"kubernetes.io/projected/0dc111dd-bc98-4dfc-8130-85c33d2271b5-kube-api-access-brs4f\") pod \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\" (UID: \"0dc111dd-bc98-4dfc-8130-85c33d2271b5\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.667055 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-utilities" (OuterVolumeSpecName: "utilities") pod "0dc111dd-bc98-4dfc-8130-85c33d2271b5" (UID: "0dc111dd-bc98-4dfc-8130-85c33d2271b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.676742 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dc111dd-bc98-4dfc-8130-85c33d2271b5-kube-api-access-brs4f" (OuterVolumeSpecName: "kube-api-access-brs4f") pod "0dc111dd-bc98-4dfc-8130-85c33d2271b5" (UID: "0dc111dd-bc98-4dfc-8130-85c33d2271b5"). InnerVolumeSpecName "kube-api-access-brs4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.719664 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dc111dd-bc98-4dfc-8130-85c33d2271b5" (UID: "0dc111dd-bc98-4dfc-8130-85c33d2271b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.761164 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-operator-metrics\") pod \"665360fe-3cbc-49bb-9316-01a6cf62174e\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.761206 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-trusted-ca\") pod \"665360fe-3cbc-49bb-9316-01a6cf62174e\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.761268 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpknx\" (UniqueName: \"kubernetes.io/projected/665360fe-3cbc-49bb-9316-01a6cf62174e-kube-api-access-fpknx\") pod \"665360fe-3cbc-49bb-9316-01a6cf62174e\" (UID: \"665360fe-3cbc-49bb-9316-01a6cf62174e\") " Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.761565 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.761582 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc111dd-bc98-4dfc-8130-85c33d2271b5-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.761591 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brs4f\" (UniqueName: \"kubernetes.io/projected/0dc111dd-bc98-4dfc-8130-85c33d2271b5-kube-api-access-brs4f\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.761849 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "665360fe-3cbc-49bb-9316-01a6cf62174e" (UID: "665360fe-3cbc-49bb-9316-01a6cf62174e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.764742 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "665360fe-3cbc-49bb-9316-01a6cf62174e" (UID: "665360fe-3cbc-49bb-9316-01a6cf62174e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.765877 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/665360fe-3cbc-49bb-9316-01a6cf62174e-kube-api-access-fpknx" (OuterVolumeSpecName: "kube-api-access-fpknx") pod "665360fe-3cbc-49bb-9316-01a6cf62174e" (UID: "665360fe-3cbc-49bb-9316-01a6cf62174e"). InnerVolumeSpecName "kube-api-access-fpknx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.862627 4957 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.862669 4957 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/665360fe-3cbc-49bb-9316-01a6cf62174e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:43 crc kubenswrapper[4957]: I1206 05:45:43.862689 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpknx\" (UniqueName: \"kubernetes.io/projected/665360fe-3cbc-49bb-9316-01a6cf62174e-kube-api-access-fpknx\") on node \"crc\" DevicePath \"\"" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.006362 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" event={"ID":"665360fe-3cbc-49bb-9316-01a6cf62174e","Type":"ContainerDied","Data":"6d6965ada9c5fe48f7b128fc5fca83a0e9ed747d5d63ee87b8db74d2468c5809"} Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.006711 4957 scope.go:117] "RemoveContainer" containerID="c71168b6de9b9d702c8a039537e6c5862ef60811a4ea53f8d23db5ef00314d83" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.006392 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gfv9x" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.012567 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44rx8" event={"ID":"0dc111dd-bc98-4dfc-8130-85c33d2271b5","Type":"ContainerDied","Data":"9ca1b4f104365c6c037684aeb9acad1fe793b27a68e4bd0b106834530fbd601e"} Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.012612 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44rx8" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.014424 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" event={"ID":"de45cbbb-4070-4df4-8b0e-c6dbe32a1a33","Type":"ContainerStarted","Data":"859d0a5baa4c16990d4199829a6f92c6d7220b3b452d33d962c41d2eaf45c9c8"} Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.014635 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.017294 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.039449 4957 scope.go:117] "RemoveContainer" containerID="b22629efd03338ca3b44ce4baab025ae3291b9cce02a093fcb29576f25a525c9" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.050136 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cbbgh" podStartSLOduration=3.050118741 podStartE2EDuration="3.050118741s" podCreationTimestamp="2025-12-06 05:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:45:44.044380871 +0000 UTC m=+383.694648503" watchObservedRunningTime="2025-12-06 05:45:44.050118741 +0000 UTC m=+383.700386373" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.066679 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-44rx8"] Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.070301 4957 scope.go:117] "RemoveContainer" containerID="ebf7b72fc6cedc0f3612c54cdeff84fb8e918f42b664f625e5d0ed1e473ad864" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.071031 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-44rx8"] Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.081857 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gfv9x"] Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.085751 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gfv9x"] Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.095177 4957 scope.go:117] "RemoveContainer" containerID="6a64e7fd6b23bcfd66f652ca8510bb86c2227cd72c2960ae989551eb68d324b0" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.670412 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" path="/var/lib/kubelet/pods/0dc111dd-bc98-4dfc-8130-85c33d2271b5/volumes" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.672196 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" path="/var/lib/kubelet/pods/5de294a0-fcab-43a3-8001-c32c3965f3f6/volumes" Dec 06 05:45:44 crc kubenswrapper[4957]: I1206 05:45:44.673676 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="665360fe-3cbc-49bb-9316-01a6cf62174e" path="/var/lib/kubelet/pods/665360fe-3cbc-49bb-9316-01a6cf62174e/volumes" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456117 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vqq25"] Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456371 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456387 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456402 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456410 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456421 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456429 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456438 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456445 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456453 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456460 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456472 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="665360fe-3cbc-49bb-9316-01a6cf62174e" containerName="marketplace-operator" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456478 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="665360fe-3cbc-49bb-9316-01a6cf62174e" containerName="marketplace-operator" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456485 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456491 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456498 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456504 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="extract-utilities" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456513 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456518 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456530 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456537 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456546 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456551 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456563 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456568 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="extract-content" Dec 06 05:45:45 crc kubenswrapper[4957]: E1206 05:45:45.456576 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456581 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456665 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de294a0-fcab-43a3-8001-c32c3965f3f6" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456673 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6be4201-af3d-4f37-a819-7cf136f1e906" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456682 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="37639dbf-8f08-428e-93e2-456e2946333f" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456689 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dc111dd-bc98-4dfc-8130-85c33d2271b5" containerName="registry-server" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.456697 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="665360fe-3cbc-49bb-9316-01a6cf62174e" containerName="marketplace-operator" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.457472 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.459305 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.461012 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vqq25"] Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.586476 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-utilities\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.586553 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-catalog-content\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.586638 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6v9p\" (UniqueName: \"kubernetes.io/projected/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-kube-api-access-t6v9p\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.653417 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8twxk"] Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.654559 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.656578 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.663180 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8twxk"] Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.688377 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6v9p\" (UniqueName: \"kubernetes.io/projected/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-kube-api-access-t6v9p\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.688443 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-utilities\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.688486 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-catalog-content\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.689159 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-catalog-content\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.689175 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-utilities\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.708431 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6v9p\" (UniqueName: \"kubernetes.io/projected/a5c2e60d-d9ec-470a-948b-70bb6184b9f4-kube-api-access-t6v9p\") pod \"redhat-marketplace-vqq25\" (UID: \"a5c2e60d-d9ec-470a-948b-70bb6184b9f4\") " pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.779038 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.789716 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g94tm\" (UniqueName: \"kubernetes.io/projected/4b08140a-aeb1-4f88-a30c-061e3ce9f403-kube-api-access-g94tm\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.790004 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b08140a-aeb1-4f88-a30c-061e3ce9f403-catalog-content\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.790161 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b08140a-aeb1-4f88-a30c-061e3ce9f403-utilities\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.891092 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b08140a-aeb1-4f88-a30c-061e3ce9f403-utilities\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.891442 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g94tm\" (UniqueName: \"kubernetes.io/projected/4b08140a-aeb1-4f88-a30c-061e3ce9f403-kube-api-access-g94tm\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.891498 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b08140a-aeb1-4f88-a30c-061e3ce9f403-catalog-content\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.891861 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b08140a-aeb1-4f88-a30c-061e3ce9f403-utilities\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.891936 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b08140a-aeb1-4f88-a30c-061e3ce9f403-catalog-content\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.908941 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g94tm\" (UniqueName: \"kubernetes.io/projected/4b08140a-aeb1-4f88-a30c-061e3ce9f403-kube-api-access-g94tm\") pod \"redhat-operators-8twxk\" (UID: \"4b08140a-aeb1-4f88-a30c-061e3ce9f403\") " pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:45 crc kubenswrapper[4957]: I1206 05:45:45.974319 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:46 crc kubenswrapper[4957]: I1206 05:45:46.185292 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vqq25"] Dec 06 05:45:46 crc kubenswrapper[4957]: W1206 05:45:46.196460 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5c2e60d_d9ec_470a_948b_70bb6184b9f4.slice/crio-a9b547db43d409ce27a8541d5c699e8c71b7d2b947386da8a1a005ebc8c81ceb WatchSource:0}: Error finding container a9b547db43d409ce27a8541d5c699e8c71b7d2b947386da8a1a005ebc8c81ceb: Status 404 returned error can't find the container with id a9b547db43d409ce27a8541d5c699e8c71b7d2b947386da8a1a005ebc8c81ceb Dec 06 05:45:46 crc kubenswrapper[4957]: I1206 05:45:46.378120 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8twxk"] Dec 06 05:45:46 crc kubenswrapper[4957]: W1206 05:45:46.387903 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b08140a_aeb1_4f88_a30c_061e3ce9f403.slice/crio-bf77a5321cd45efff9b27480f8bccabb245e7c28a571b01bd10bf034e636b026 WatchSource:0}: Error finding container bf77a5321cd45efff9b27480f8bccabb245e7c28a571b01bd10bf034e636b026: Status 404 returned error can't find the container with id bf77a5321cd45efff9b27480f8bccabb245e7c28a571b01bd10bf034e636b026 Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.034071 4957 generic.go:334] "Generic (PLEG): container finished" podID="4b08140a-aeb1-4f88-a30c-061e3ce9f403" containerID="855429bbe5e476831000b6f7c748b578f7f8dd9628cee0e853b1bb361186270c" exitCode=0 Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.034153 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8twxk" event={"ID":"4b08140a-aeb1-4f88-a30c-061e3ce9f403","Type":"ContainerDied","Data":"855429bbe5e476831000b6f7c748b578f7f8dd9628cee0e853b1bb361186270c"} Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.034192 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8twxk" event={"ID":"4b08140a-aeb1-4f88-a30c-061e3ce9f403","Type":"ContainerStarted","Data":"bf77a5321cd45efff9b27480f8bccabb245e7c28a571b01bd10bf034e636b026"} Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.039402 4957 generic.go:334] "Generic (PLEG): container finished" podID="a5c2e60d-d9ec-470a-948b-70bb6184b9f4" containerID="e4cd36d84b2ff7a6b650154392178ac52292627e3434fe15a335de44226b42f8" exitCode=0 Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.039448 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqq25" event={"ID":"a5c2e60d-d9ec-470a-948b-70bb6184b9f4","Type":"ContainerDied","Data":"e4cd36d84b2ff7a6b650154392178ac52292627e3434fe15a335de44226b42f8"} Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.039475 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqq25" event={"ID":"a5c2e60d-d9ec-470a-948b-70bb6184b9f4","Type":"ContainerStarted","Data":"a9b547db43d409ce27a8541d5c699e8c71b7d2b947386da8a1a005ebc8c81ceb"} Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.853467 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-crj2z"] Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.854943 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.857199 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.863001 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crj2z"] Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.922785 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxn8w\" (UniqueName: \"kubernetes.io/projected/5be927bc-2bb0-4676-929c-bfd93bacc0f5-kube-api-access-nxn8w\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.922859 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be927bc-2bb0-4676-929c-bfd93bacc0f5-utilities\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:47 crc kubenswrapper[4957]: I1206 05:45:47.922884 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be927bc-2bb0-4676-929c-bfd93bacc0f5-catalog-content\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.024607 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be927bc-2bb0-4676-929c-bfd93bacc0f5-utilities\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.024694 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be927bc-2bb0-4676-929c-bfd93bacc0f5-catalog-content\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.024817 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxn8w\" (UniqueName: \"kubernetes.io/projected/5be927bc-2bb0-4676-929c-bfd93bacc0f5-kube-api-access-nxn8w\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.025465 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be927bc-2bb0-4676-929c-bfd93bacc0f5-utilities\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.025746 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be927bc-2bb0-4676-929c-bfd93bacc0f5-catalog-content\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.049638 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxn8w\" (UniqueName: \"kubernetes.io/projected/5be927bc-2bb0-4676-929c-bfd93bacc0f5-kube-api-access-nxn8w\") pod \"certified-operators-crj2z\" (UID: \"5be927bc-2bb0-4676-929c-bfd93bacc0f5\") " pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.056276 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wmqv2"] Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.057507 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.060233 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.066117 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wmqv2"] Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.125569 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-utilities\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.125620 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-catalog-content\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.125695 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pshk9\" (UniqueName: \"kubernetes.io/projected/6a8ca263-88ab-45bb-80d4-70add07a7b14-kube-api-access-pshk9\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.194001 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.227141 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-utilities\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.227258 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-catalog-content\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.227450 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pshk9\" (UniqueName: \"kubernetes.io/projected/6a8ca263-88ab-45bb-80d4-70add07a7b14-kube-api-access-pshk9\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.227924 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-utilities\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.228212 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-catalog-content\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.247377 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pshk9\" (UniqueName: \"kubernetes.io/projected/6a8ca263-88ab-45bb-80d4-70add07a7b14-kube-api-access-pshk9\") pod \"community-operators-wmqv2\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.398452 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.464307 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crj2z"] Dec 06 05:45:48 crc kubenswrapper[4957]: W1206 05:45:48.483761 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5be927bc_2bb0_4676_929c_bfd93bacc0f5.slice/crio-c3945d8d16b64f9c1a3770583cb9e80b3ee67ab81151d1217dd330314181beb0 WatchSource:0}: Error finding container c3945d8d16b64f9c1a3770583cb9e80b3ee67ab81151d1217dd330314181beb0: Status 404 returned error can't find the container with id c3945d8d16b64f9c1a3770583cb9e80b3ee67ab81151d1217dd330314181beb0 Dec 06 05:45:48 crc kubenswrapper[4957]: I1206 05:45:48.804324 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wmqv2"] Dec 06 05:45:48 crc kubenswrapper[4957]: W1206 05:45:48.813497 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a8ca263_88ab_45bb_80d4_70add07a7b14.slice/crio-e48c441b1f04072e291c583ece35897528c3336ec1f8475a6fec304e36f0aa84 WatchSource:0}: Error finding container e48c441b1f04072e291c583ece35897528c3336ec1f8475a6fec304e36f0aa84: Status 404 returned error can't find the container with id e48c441b1f04072e291c583ece35897528c3336ec1f8475a6fec304e36f0aa84 Dec 06 05:45:49 crc kubenswrapper[4957]: I1206 05:45:49.056868 4957 generic.go:334] "Generic (PLEG): container finished" podID="5be927bc-2bb0-4676-929c-bfd93bacc0f5" containerID="08b6df5f0141f4aeb3a2cc378950b48b324237095592a8a8eb2089a32e84e07a" exitCode=0 Dec 06 05:45:49 crc kubenswrapper[4957]: I1206 05:45:49.056955 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crj2z" event={"ID":"5be927bc-2bb0-4676-929c-bfd93bacc0f5","Type":"ContainerDied","Data":"08b6df5f0141f4aeb3a2cc378950b48b324237095592a8a8eb2089a32e84e07a"} Dec 06 05:45:49 crc kubenswrapper[4957]: I1206 05:45:49.057065 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crj2z" event={"ID":"5be927bc-2bb0-4676-929c-bfd93bacc0f5","Type":"ContainerStarted","Data":"c3945d8d16b64f9c1a3770583cb9e80b3ee67ab81151d1217dd330314181beb0"} Dec 06 05:45:49 crc kubenswrapper[4957]: I1206 05:45:49.059513 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmqv2" event={"ID":"6a8ca263-88ab-45bb-80d4-70add07a7b14","Type":"ContainerStarted","Data":"305fd7f55e6d3ec6aa50dfb8f828124b00513db55ab707201233f4f13b224087"} Dec 06 05:45:49 crc kubenswrapper[4957]: I1206 05:45:49.059569 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmqv2" event={"ID":"6a8ca263-88ab-45bb-80d4-70add07a7b14","Type":"ContainerStarted","Data":"e48c441b1f04072e291c583ece35897528c3336ec1f8475a6fec304e36f0aa84"} Dec 06 05:45:50 crc kubenswrapper[4957]: I1206 05:45:50.068124 4957 generic.go:334] "Generic (PLEG): container finished" podID="4b08140a-aeb1-4f88-a30c-061e3ce9f403" containerID="28fffc841582df904c951ed34b59a6860739a98d7410569b6d64f7aab9e4d8e5" exitCode=0 Dec 06 05:45:50 crc kubenswrapper[4957]: I1206 05:45:50.068331 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8twxk" event={"ID":"4b08140a-aeb1-4f88-a30c-061e3ce9f403","Type":"ContainerDied","Data":"28fffc841582df904c951ed34b59a6860739a98d7410569b6d64f7aab9e4d8e5"} Dec 06 05:45:50 crc kubenswrapper[4957]: I1206 05:45:50.075720 4957 generic.go:334] "Generic (PLEG): container finished" podID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerID="305fd7f55e6d3ec6aa50dfb8f828124b00513db55ab707201233f4f13b224087" exitCode=0 Dec 06 05:45:50 crc kubenswrapper[4957]: I1206 05:45:50.076041 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmqv2" event={"ID":"6a8ca263-88ab-45bb-80d4-70add07a7b14","Type":"ContainerDied","Data":"305fd7f55e6d3ec6aa50dfb8f828124b00513db55ab707201233f4f13b224087"} Dec 06 05:45:50 crc kubenswrapper[4957]: I1206 05:45:50.079370 4957 generic.go:334] "Generic (PLEG): container finished" podID="a5c2e60d-d9ec-470a-948b-70bb6184b9f4" containerID="c6b65b3a01dda61f754d7626061061cc536977577263a3ca787890c4ef48e161" exitCode=0 Dec 06 05:45:50 crc kubenswrapper[4957]: I1206 05:45:50.079541 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqq25" event={"ID":"a5c2e60d-d9ec-470a-948b-70bb6184b9f4","Type":"ContainerDied","Data":"c6b65b3a01dda61f754d7626061061cc536977577263a3ca787890c4ef48e161"} Dec 06 05:45:51 crc kubenswrapper[4957]: I1206 05:45:51.093983 4957 generic.go:334] "Generic (PLEG): container finished" podID="5be927bc-2bb0-4676-929c-bfd93bacc0f5" containerID="3148dd2003e55115abf4b39d4999c5fbf4815cd63e3f4f9d6964f87ca4a6275f" exitCode=0 Dec 06 05:45:51 crc kubenswrapper[4957]: I1206 05:45:51.094084 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crj2z" event={"ID":"5be927bc-2bb0-4676-929c-bfd93bacc0f5","Type":"ContainerDied","Data":"3148dd2003e55115abf4b39d4999c5fbf4815cd63e3f4f9d6964f87ca4a6275f"} Dec 06 05:45:52 crc kubenswrapper[4957]: I1206 05:45:52.100542 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqq25" event={"ID":"a5c2e60d-d9ec-470a-948b-70bb6184b9f4","Type":"ContainerStarted","Data":"737080c944b7e1778725cf8b41340f4e8c2ec176db5080f0d78c4f0f9582b19a"} Dec 06 05:45:52 crc kubenswrapper[4957]: I1206 05:45:52.116851 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vqq25" podStartSLOduration=2.966501302 podStartE2EDuration="7.116819015s" podCreationTimestamp="2025-12-06 05:45:45 +0000 UTC" firstStartedPulling="2025-12-06 05:45:47.04207855 +0000 UTC m=+386.692346182" lastFinishedPulling="2025-12-06 05:45:51.192396263 +0000 UTC m=+390.842663895" observedRunningTime="2025-12-06 05:45:52.116555648 +0000 UTC m=+391.766823280" watchObservedRunningTime="2025-12-06 05:45:52.116819015 +0000 UTC m=+391.767086647" Dec 06 05:45:53 crc kubenswrapper[4957]: I1206 05:45:53.108001 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8twxk" event={"ID":"4b08140a-aeb1-4f88-a30c-061e3ce9f403","Type":"ContainerStarted","Data":"62c8fb23dbd830d300cea8504f23a7a7e431f09be75ea733599c93f3bb463748"} Dec 06 05:45:53 crc kubenswrapper[4957]: I1206 05:45:53.109897 4957 generic.go:334] "Generic (PLEG): container finished" podID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerID="04a719f3864b72821b857ef55dc6d96f190ab45597c647381b46c1d052f8dbf7" exitCode=0 Dec 06 05:45:53 crc kubenswrapper[4957]: I1206 05:45:53.109963 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmqv2" event={"ID":"6a8ca263-88ab-45bb-80d4-70add07a7b14","Type":"ContainerDied","Data":"04a719f3864b72821b857ef55dc6d96f190ab45597c647381b46c1d052f8dbf7"} Dec 06 05:45:53 crc kubenswrapper[4957]: I1206 05:45:53.112872 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crj2z" event={"ID":"5be927bc-2bb0-4676-929c-bfd93bacc0f5","Type":"ContainerStarted","Data":"b4f53b19e8e5c3b37000605e61eaea867323ae935a305e511e14c95f14d7b7eb"} Dec 06 05:45:53 crc kubenswrapper[4957]: I1206 05:45:53.127391 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8twxk" podStartSLOduration=3.6115338120000002 podStartE2EDuration="8.12737474s" podCreationTimestamp="2025-12-06 05:45:45 +0000 UTC" firstStartedPulling="2025-12-06 05:45:47.037238894 +0000 UTC m=+386.687506526" lastFinishedPulling="2025-12-06 05:45:51.553079822 +0000 UTC m=+391.203347454" observedRunningTime="2025-12-06 05:45:53.124913952 +0000 UTC m=+392.775181584" watchObservedRunningTime="2025-12-06 05:45:53.12737474 +0000 UTC m=+392.777642372" Dec 06 05:45:53 crc kubenswrapper[4957]: I1206 05:45:53.177838 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-crj2z" podStartSLOduration=3.200383417 podStartE2EDuration="6.177800714s" podCreationTimestamp="2025-12-06 05:45:47 +0000 UTC" firstStartedPulling="2025-12-06 05:45:49.098060006 +0000 UTC m=+388.748327638" lastFinishedPulling="2025-12-06 05:45:52.075477303 +0000 UTC m=+391.725744935" observedRunningTime="2025-12-06 05:45:53.162095541 +0000 UTC m=+392.812363173" watchObservedRunningTime="2025-12-06 05:45:53.177800714 +0000 UTC m=+392.828068346" Dec 06 05:45:55 crc kubenswrapper[4957]: I1206 05:45:55.129188 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmqv2" event={"ID":"6a8ca263-88ab-45bb-80d4-70add07a7b14","Type":"ContainerStarted","Data":"08045462e07788e92a2c6ba18daf7194e7416e1a373b015b135233db78000288"} Dec 06 05:45:55 crc kubenswrapper[4957]: I1206 05:45:55.779617 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:55 crc kubenswrapper[4957]: I1206 05:45:55.779787 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:55 crc kubenswrapper[4957]: I1206 05:45:55.829675 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:55 crc kubenswrapper[4957]: I1206 05:45:55.974613 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:55 crc kubenswrapper[4957]: I1206 05:45:55.974656 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:45:56 crc kubenswrapper[4957]: I1206 05:45:56.166588 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wmqv2" podStartSLOduration=4.146824333 podStartE2EDuration="8.166561524s" podCreationTimestamp="2025-12-06 05:45:48 +0000 UTC" firstStartedPulling="2025-12-06 05:45:50.077405767 +0000 UTC m=+389.727673399" lastFinishedPulling="2025-12-06 05:45:54.097142968 +0000 UTC m=+393.747410590" observedRunningTime="2025-12-06 05:45:56.164086915 +0000 UTC m=+395.814354547" watchObservedRunningTime="2025-12-06 05:45:56.166561524 +0000 UTC m=+395.816829146" Dec 06 05:45:56 crc kubenswrapper[4957]: I1206 05:45:56.186057 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vqq25" Dec 06 05:45:57 crc kubenswrapper[4957]: I1206 05:45:57.016026 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8twxk" podUID="4b08140a-aeb1-4f88-a30c-061e3ce9f403" containerName="registry-server" probeResult="failure" output=< Dec 06 05:45:57 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 05:45:57 crc kubenswrapper[4957]: > Dec 06 05:45:58 crc kubenswrapper[4957]: I1206 05:45:58.194391 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:58 crc kubenswrapper[4957]: I1206 05:45:58.194443 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:58 crc kubenswrapper[4957]: I1206 05:45:58.234785 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:58 crc kubenswrapper[4957]: I1206 05:45:58.399710 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:58 crc kubenswrapper[4957]: I1206 05:45:58.400037 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:58 crc kubenswrapper[4957]: I1206 05:45:58.466507 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:45:59 crc kubenswrapper[4957]: I1206 05:45:59.214021 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-crj2z" Dec 06 05:45:59 crc kubenswrapper[4957]: I1206 05:45:59.222918 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:46:06 crc kubenswrapper[4957]: I1206 05:46:06.042135 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:46:06 crc kubenswrapper[4957]: I1206 05:46:06.098987 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8twxk" Dec 06 05:46:08 crc kubenswrapper[4957]: I1206 05:46:08.680268 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" podUID="8c269452-c147-4adc-9280-7dd91cff2687" containerName="registry" containerID="cri-o://386075d3ff234a92cadf00ef8864691d2f2c9ed3aa7012520e0111ced9c791aa" gracePeriod=30 Dec 06 05:46:10 crc kubenswrapper[4957]: I1206 05:46:10.180690 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:46:10 crc kubenswrapper[4957]: I1206 05:46:10.181329 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:46:10 crc kubenswrapper[4957]: I1206 05:46:10.181402 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:46:10 crc kubenswrapper[4957]: I1206 05:46:10.182562 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ffb51b23be1ebfd26c70f252394c0ca06f92ba61d22e53e094223c893ca1069"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:46:10 crc kubenswrapper[4957]: I1206 05:46:10.182632 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://7ffb51b23be1ebfd26c70f252394c0ca06f92ba61d22e53e094223c893ca1069" gracePeriod=600 Dec 06 05:46:12 crc kubenswrapper[4957]: I1206 05:46:12.242245 4957 generic.go:334] "Generic (PLEG): container finished" podID="8c269452-c147-4adc-9280-7dd91cff2687" containerID="386075d3ff234a92cadf00ef8864691d2f2c9ed3aa7012520e0111ced9c791aa" exitCode=0 Dec 06 05:46:12 crc kubenswrapper[4957]: I1206 05:46:12.242341 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" event={"ID":"8c269452-c147-4adc-9280-7dd91cff2687","Type":"ContainerDied","Data":"386075d3ff234a92cadf00ef8864691d2f2c9ed3aa7012520e0111ced9c791aa"} Dec 06 05:46:12 crc kubenswrapper[4957]: I1206 05:46:12.245185 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="7ffb51b23be1ebfd26c70f252394c0ca06f92ba61d22e53e094223c893ca1069" exitCode=0 Dec 06 05:46:12 crc kubenswrapper[4957]: I1206 05:46:12.245252 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"7ffb51b23be1ebfd26c70f252394c0ca06f92ba61d22e53e094223c893ca1069"} Dec 06 05:46:12 crc kubenswrapper[4957]: I1206 05:46:12.245313 4957 scope.go:117] "RemoveContainer" containerID="decdee56891dd06d4debf1c64b7dfa19183168d6cfcda6d14b3ef8b361c806e3" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.132185 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.261182 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" event={"ID":"8c269452-c147-4adc-9280-7dd91cff2687","Type":"ContainerDied","Data":"171d583a4ba9f746d1290bffdba8fb0ca8bc3ecb0dba69eef857fd88b366cfea"} Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.261488 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xlz55" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.261493 4957 scope.go:117] "RemoveContainer" containerID="386075d3ff234a92cadf00ef8864691d2f2c9ed3aa7012520e0111ced9c791aa" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.265768 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"af5d9f710568e718f75a967bf395611eae1b65dcacd5acd1a647b8480dd1df56"} Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.284415 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qkw4\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-kube-api-access-6qkw4\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.284484 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-bound-sa-token\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.284523 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c269452-c147-4adc-9280-7dd91cff2687-installation-pull-secrets\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.284593 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-registry-certificates\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.284636 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-registry-tls\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.284987 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.285021 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c269452-c147-4adc-9280-7dd91cff2687-ca-trust-extracted\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.285087 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-trusted-ca\") pod \"8c269452-c147-4adc-9280-7dd91cff2687\" (UID: \"8c269452-c147-4adc-9280-7dd91cff2687\") " Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.286384 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.287865 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.293572 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.294179 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c269452-c147-4adc-9280-7dd91cff2687-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.297412 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.298124 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-kube-api-access-6qkw4" (OuterVolumeSpecName: "kube-api-access-6qkw4") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "kube-api-access-6qkw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.299575 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.306826 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c269452-c147-4adc-9280-7dd91cff2687-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8c269452-c147-4adc-9280-7dd91cff2687" (UID: "8c269452-c147-4adc-9280-7dd91cff2687"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.387383 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qkw4\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-kube-api-access-6qkw4\") on node \"crc\" DevicePath \"\"" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.387438 4957 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.387455 4957 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c269452-c147-4adc-9280-7dd91cff2687-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.387468 4957 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.387482 4957 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c269452-c147-4adc-9280-7dd91cff2687-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.387495 4957 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c269452-c147-4adc-9280-7dd91cff2687-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.387504 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c269452-c147-4adc-9280-7dd91cff2687-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.614241 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xlz55"] Dec 06 05:46:13 crc kubenswrapper[4957]: I1206 05:46:13.618966 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xlz55"] Dec 06 05:46:14 crc kubenswrapper[4957]: I1206 05:46:14.676073 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c269452-c147-4adc-9280-7dd91cff2687" path="/var/lib/kubelet/pods/8c269452-c147-4adc-9280-7dd91cff2687/volumes" Dec 06 05:48:20 crc kubenswrapper[4957]: I1206 05:48:20.902007 4957 scope.go:117] "RemoveContainer" containerID="f7a11a3d0bcbbea2a8b00ebd1141065f433835fb4f0c621d4e06a1819922961d" Dec 06 05:48:20 crc kubenswrapper[4957]: I1206 05:48:20.919815 4957 scope.go:117] "RemoveContainer" containerID="c6f95c6ceebca8842826a5a7327239d6fdf3993ce68f9b84321687180efd4b6b" Dec 06 05:48:20 crc kubenswrapper[4957]: I1206 05:48:20.950337 4957 scope.go:117] "RemoveContainer" containerID="bfd867228f5dd65e3fbc69a7480136147f047364dc11e4fa4dde4b962cee4468" Dec 06 05:48:40 crc kubenswrapper[4957]: I1206 05:48:40.180659 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:48:40 crc kubenswrapper[4957]: I1206 05:48:40.181366 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:49:10 crc kubenswrapper[4957]: I1206 05:49:10.180742 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:49:10 crc kubenswrapper[4957]: I1206 05:49:10.181642 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:49:20 crc kubenswrapper[4957]: I1206 05:49:20.985558 4957 scope.go:117] "RemoveContainer" containerID="94336eb85ebac1c144be1e19606b1b7792113501120787462d880dc6d4f1bbba" Dec 06 05:49:21 crc kubenswrapper[4957]: I1206 05:49:21.012540 4957 scope.go:117] "RemoveContainer" containerID="0c575c01c3a70c5ac7e74ad250d4fad45693a81ef99c329541f89de52d88e316" Dec 06 05:49:21 crc kubenswrapper[4957]: I1206 05:49:21.052252 4957 scope.go:117] "RemoveContainer" containerID="1b9cc0fb0f4c20a4b4d6795c4099fa80416d82528d8d2b5d6857018802b0f973" Dec 06 05:49:21 crc kubenswrapper[4957]: I1206 05:49:21.068418 4957 scope.go:117] "RemoveContainer" containerID="d24ce9a9584511200ea542b47e2b283de7e53ca54f0cff5ce91551212750ff3e" Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.181494 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.182180 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.182228 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.182817 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af5d9f710568e718f75a967bf395611eae1b65dcacd5acd1a647b8480dd1df56"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.182895 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://af5d9f710568e718f75a967bf395611eae1b65dcacd5acd1a647b8480dd1df56" gracePeriod=600 Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.789464 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="af5d9f710568e718f75a967bf395611eae1b65dcacd5acd1a647b8480dd1df56" exitCode=0 Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.789541 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"af5d9f710568e718f75a967bf395611eae1b65dcacd5acd1a647b8480dd1df56"} Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.790053 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"0b0c4f58bdcf5290181ec1a457cbeb02772611a1bd68d33a39e309f1aaee0927"} Dec 06 05:49:40 crc kubenswrapper[4957]: I1206 05:49:40.790088 4957 scope.go:117] "RemoveContainer" containerID="7ffb51b23be1ebfd26c70f252394c0ca06f92ba61d22e53e094223c893ca1069" Dec 06 05:51:40 crc kubenswrapper[4957]: I1206 05:51:40.180820 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:51:40 crc kubenswrapper[4957]: I1206 05:51:40.181644 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:51:52 crc kubenswrapper[4957]: I1206 05:51:52.677679 4957 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.554816 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b256m"] Dec 06 05:52:00 crc kubenswrapper[4957]: E1206 05:52:00.556148 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c269452-c147-4adc-9280-7dd91cff2687" containerName="registry" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.556168 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c269452-c147-4adc-9280-7dd91cff2687" containerName="registry" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.556295 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c269452-c147-4adc-9280-7dd91cff2687" containerName="registry" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.556935 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.561662 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.562517 4957 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-vw48m" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.563900 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.577474 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6557k"] Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.578438 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.581857 4957 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qczdm" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.598935 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-59bbt"] Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.600252 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-59bbt" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.605437 4957 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-dhb2v" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.613995 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-59bbt"] Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.631264 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b256m"] Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.634254 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6557k"] Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.702687 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msgcn\" (UniqueName: \"kubernetes.io/projected/ff08ac61-0abb-4f45-a647-57aca5172e36-kube-api-access-msgcn\") pod \"cert-manager-5b446d88c5-59bbt\" (UID: \"ff08ac61-0abb-4f45-a647-57aca5172e36\") " pod="cert-manager/cert-manager-5b446d88c5-59bbt" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.702985 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwf78\" (UniqueName: \"kubernetes.io/projected/14f39aaa-5904-4a8d-86ad-ac487456788b-kube-api-access-zwf78\") pod \"cert-manager-cainjector-7f985d654d-6557k\" (UID: \"14f39aaa-5904-4a8d-86ad-ac487456788b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.703093 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4zj8\" (UniqueName: \"kubernetes.io/projected/8ebad773-0957-4bb8-b583-a0cff091ff46-kube-api-access-p4zj8\") pod \"cert-manager-webhook-5655c58dd6-b256m\" (UID: \"8ebad773-0957-4bb8-b583-a0cff091ff46\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.804689 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msgcn\" (UniqueName: \"kubernetes.io/projected/ff08ac61-0abb-4f45-a647-57aca5172e36-kube-api-access-msgcn\") pod \"cert-manager-5b446d88c5-59bbt\" (UID: \"ff08ac61-0abb-4f45-a647-57aca5172e36\") " pod="cert-manager/cert-manager-5b446d88c5-59bbt" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.805228 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwf78\" (UniqueName: \"kubernetes.io/projected/14f39aaa-5904-4a8d-86ad-ac487456788b-kube-api-access-zwf78\") pod \"cert-manager-cainjector-7f985d654d-6557k\" (UID: \"14f39aaa-5904-4a8d-86ad-ac487456788b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.805425 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4zj8\" (UniqueName: \"kubernetes.io/projected/8ebad773-0957-4bb8-b583-a0cff091ff46-kube-api-access-p4zj8\") pod \"cert-manager-webhook-5655c58dd6-b256m\" (UID: \"8ebad773-0957-4bb8-b583-a0cff091ff46\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.826314 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4zj8\" (UniqueName: \"kubernetes.io/projected/8ebad773-0957-4bb8-b583-a0cff091ff46-kube-api-access-p4zj8\") pod \"cert-manager-webhook-5655c58dd6-b256m\" (UID: \"8ebad773-0957-4bb8-b583-a0cff091ff46\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.826418 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msgcn\" (UniqueName: \"kubernetes.io/projected/ff08ac61-0abb-4f45-a647-57aca5172e36-kube-api-access-msgcn\") pod \"cert-manager-5b446d88c5-59bbt\" (UID: \"ff08ac61-0abb-4f45-a647-57aca5172e36\") " pod="cert-manager/cert-manager-5b446d88c5-59bbt" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.830989 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwf78\" (UniqueName: \"kubernetes.io/projected/14f39aaa-5904-4a8d-86ad-ac487456788b-kube-api-access-zwf78\") pod \"cert-manager-cainjector-7f985d654d-6557k\" (UID: \"14f39aaa-5904-4a8d-86ad-ac487456788b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.876356 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.898364 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" Dec 06 05:52:00 crc kubenswrapper[4957]: I1206 05:52:00.920291 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-59bbt" Dec 06 05:52:01 crc kubenswrapper[4957]: I1206 05:52:01.114587 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6557k"] Dec 06 05:52:01 crc kubenswrapper[4957]: I1206 05:52:01.125235 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 05:52:01 crc kubenswrapper[4957]: I1206 05:52:01.145688 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b256m"] Dec 06 05:52:01 crc kubenswrapper[4957]: I1206 05:52:01.185066 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-59bbt"] Dec 06 05:52:01 crc kubenswrapper[4957]: I1206 05:52:01.654275 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" event={"ID":"8ebad773-0957-4bb8-b583-a0cff091ff46","Type":"ContainerStarted","Data":"eb6bc14617af34b0a3dd20826ddeb1c08d672dcfe07ff95c33acd96bf1d75d47"} Dec 06 05:52:01 crc kubenswrapper[4957]: I1206 05:52:01.655591 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" event={"ID":"14f39aaa-5904-4a8d-86ad-ac487456788b","Type":"ContainerStarted","Data":"03dbaf5040b42311b4863a0123d3a14417e00b200e8a5f682b4108ef498d80e0"} Dec 06 05:52:01 crc kubenswrapper[4957]: I1206 05:52:01.657092 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-59bbt" event={"ID":"ff08ac61-0abb-4f45-a647-57aca5172e36","Type":"ContainerStarted","Data":"40efb7e55f975366db11b9bbcae4a4ea3584c9e088e8f8780141cd3f89298387"} Dec 06 05:52:10 crc kubenswrapper[4957]: I1206 05:52:10.181140 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:52:10 crc kubenswrapper[4957]: I1206 05:52:10.181615 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.117955 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hlm8s"] Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.130576 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-controller" containerID="cri-o://dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.131086 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="sbdb" containerID="cri-o://1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.131151 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="nbdb" containerID="cri-o://526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.131199 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="northd" containerID="cri-o://fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.131243 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.131288 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-node" containerID="cri-o://483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.131332 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-acl-logging" containerID="cri-o://30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.213644 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" containerID="cri-o://de3f99007b23a6f713f546b61089c454e66fbd1f9502857afebc44f1e5f7f29d" gracePeriod=30 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.719768 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/3.log" Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.728544 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-acl-logging/0.log" Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.729349 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4" exitCode=143 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.729406 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4"} Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.731444 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/2.log" Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.731820 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/1.log" Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.731872 4957 generic.go:334] "Generic (PLEG): container finished" podID="1384ce42-5e9a-4d6a-a299-f564d1494b0a" containerID="d2f2175513b8ccbbbf889c9311445008e47014dd16395282f88be03d4396421e" exitCode=2 Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.731911 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerDied","Data":"d2f2175513b8ccbbbf889c9311445008e47014dd16395282f88be03d4396421e"} Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.731976 4957 scope.go:117] "RemoveContainer" containerID="9f96900527015396e7ec641f466b21a426e8f981d5890e7561c584b0014a3f41" Dec 06 05:52:11 crc kubenswrapper[4957]: I1206 05:52:11.732847 4957 scope.go:117] "RemoveContainer" containerID="d2f2175513b8ccbbbf889c9311445008e47014dd16395282f88be03d4396421e" Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.742090 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovnkube-controller/3.log" Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.745826 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-acl-logging/0.log" Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.746559 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-controller/0.log" Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747102 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="de3f99007b23a6f713f546b61089c454e66fbd1f9502857afebc44f1e5f7f29d" exitCode=0 Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747126 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc" exitCode=0 Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747134 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384" exitCode=0 Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747143 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe" exitCode=0 Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747152 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b" exitCode=0 Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747160 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d" exitCode=0 Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747167 4957 generic.go:334] "Generic (PLEG): container finished" podID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerID="dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f" exitCode=143 Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747180 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"de3f99007b23a6f713f546b61089c454e66fbd1f9502857afebc44f1e5f7f29d"} Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747238 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc"} Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747252 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384"} Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747263 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe"} Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747278 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b"} Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747288 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d"} Dec 06 05:52:12 crc kubenswrapper[4957]: I1206 05:52:12.747296 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f"} Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.109975 4957 scope.go:117] "RemoveContainer" containerID="4842887b25af82f6efb7dd6a9bc359d51ea12ecc78cef1ae10f3fcd2f03f1e16" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.134573 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-acl-logging/0.log" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.135125 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-controller/0.log" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.135665 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.198522 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-config\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.198698 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-var-lib-openvswitch\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.199983 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-ovn-kubernetes\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.199023 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.199292 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200094 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-node-log" (OuterVolumeSpecName: "node-log") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200061 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-node-log\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200037 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200184 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bpjx\" (UniqueName: \"kubernetes.io/projected/fa635ecb-5324-449e-a8f3-7a9bfdca7064-kube-api-access-6bpjx\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200222 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovn-node-metrics-cert\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200249 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-bin\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200288 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-kubelet\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200331 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-script-lib\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200354 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-env-overrides\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200374 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-var-lib-cni-networks-ovn-kubernetes\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200400 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-systemd\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200418 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-ovn\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200435 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-etc-openvswitch\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200438 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nwk7m"] Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200591 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-netns\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200625 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-systemd-units\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200644 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-slash\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200667 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-log-socket\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200682 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-netd\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200756 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-openvswitch\") pod \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\" (UID: \"fa635ecb-5324-449e-a8f3-7a9bfdca7064\") " Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200776 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200793 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200803 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kubecfg-setup" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200811 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kubecfg-setup" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200826 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200879 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200888 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="northd" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200894 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="northd" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200905 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-acl-logging" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200911 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-acl-logging" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200921 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200929 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200943 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200950 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200960 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="nbdb" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200966 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="nbdb" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200978 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-node" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.200987 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-node" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.200997 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201005 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.201014 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="sbdb" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201020 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="sbdb" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201117 4957 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201131 4957 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201140 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201155 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-node" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201166 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="nbdb" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201176 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201186 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-acl-logging" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201195 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovn-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201205 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="sbdb" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201213 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201220 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201227 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="northd" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201233 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.201343 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201351 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: E1206 05:52:13.201364 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201371 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201480 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" containerName="ovnkube-controller" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201141 4957 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201941 4957 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201164 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201865 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201885 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201937 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201968 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.201997 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-log-socket" (OuterVolumeSpecName: "log-socket") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.202024 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.202029 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-slash" (OuterVolumeSpecName: "host-slash") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.202043 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.202036 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.202067 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.202533 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.202640 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.204100 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.214536 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa635ecb-5324-449e-a8f3-7a9bfdca7064-kube-api-access-6bpjx" (OuterVolumeSpecName: "kube-api-access-6bpjx") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "kube-api-access-6bpjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.214917 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.219978 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "fa635ecb-5324-449e-a8f3-7a9bfdca7064" (UID: "fa635ecb-5324-449e-a8f3-7a9bfdca7064"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304567 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-cni-netd\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304633 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-systemd\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304656 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovnkube-config\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304682 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-env-overrides\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304703 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovn-node-metrics-cert\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304732 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-run-ovn-kubernetes\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304765 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovnkube-script-lib\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304796 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-systemd-units\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304818 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-run-netns\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304874 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-var-lib-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304900 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-log-socket\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304928 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304963 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.304989 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-ovn\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305010 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-node-log\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305037 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-slash\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305086 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-etc-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305106 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-cni-bin\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305124 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fddpk\" (UniqueName: \"kubernetes.io/projected/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-kube-api-access-fddpk\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305145 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-kubelet\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305225 4957 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305243 4957 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305255 4957 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa635ecb-5324-449e-a8f3-7a9bfdca7064-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305266 4957 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305277 4957 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305290 4957 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305301 4957 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305312 4957 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305323 4957 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305333 4957 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305342 4957 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305351 4957 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305361 4957 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305372 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bpjx\" (UniqueName: \"kubernetes.io/projected/fa635ecb-5324-449e-a8f3-7a9bfdca7064-kube-api-access-6bpjx\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305382 4957 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa635ecb-5324-449e-a8f3-7a9bfdca7064-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.305391 4957 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa635ecb-5324-449e-a8f3-7a9bfdca7064-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406617 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-systemd-units\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406665 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-run-netns\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406686 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-var-lib-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406710 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-log-socket\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406729 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406885 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-run-netns\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406893 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-log-socket\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406932 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-systemd-units\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406947 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406946 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-var-lib-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.407046 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.406779 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.407999 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-ovn\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408027 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-node-log\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408063 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-slash\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408074 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-ovn\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408134 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-node-log\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408165 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-slash\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408202 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-etc-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408227 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-cni-bin\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408245 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fddpk\" (UniqueName: \"kubernetes.io/projected/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-kube-api-access-fddpk\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408265 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-kubelet\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408304 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-cni-netd\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408331 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-systemd\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408349 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovnkube-config\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408385 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-env-overrides\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408415 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovn-node-metrics-cert\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408450 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-run-ovn-kubernetes\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408502 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovnkube-script-lib\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.408994 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-cni-netd\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.409051 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-etc-openvswitch\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.409082 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-cni-bin\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.409574 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovnkube-script-lib\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.409600 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-kubelet\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.409642 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-run-systemd\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.409642 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-env-overrides\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.409693 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-host-run-ovn-kubernetes\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.410247 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovnkube-config\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.421036 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-ovn-node-metrics-cert\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.439556 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fddpk\" (UniqueName: \"kubernetes.io/projected/82e76364-26d3-4cb8-b57f-c8f03ce09a7a-kube-api-access-fddpk\") pod \"ovnkube-node-nwk7m\" (UID: \"82e76364-26d3-4cb8-b57f-c8f03ce09a7a\") " pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.566824 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.763359 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-acl-logging/0.log" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.763915 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-controller/0.log" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.764370 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" event={"ID":"fa635ecb-5324-449e-a8f3-7a9bfdca7064","Type":"ContainerDied","Data":"f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5"} Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.764460 4957 scope.go:117] "RemoveContainer" containerID="de3f99007b23a6f713f546b61089c454e66fbd1f9502857afebc44f1e5f7f29d" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.764466 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hlm8s" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.766388 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/2.log" Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.823060 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hlm8s"] Dec 06 05:52:13 crc kubenswrapper[4957]: I1206 05:52:13.828164 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hlm8s"] Dec 06 05:52:14 crc kubenswrapper[4957]: I1206 05:52:14.672192 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa635ecb-5324-449e-a8f3-7a9bfdca7064" path="/var/lib/kubelet/pods/fa635ecb-5324-449e-a8f3-7a9bfdca7064/volumes" Dec 06 05:52:14 crc kubenswrapper[4957]: I1206 05:52:14.775535 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-69r8l_1384ce42-5e9a-4d6a-a299-f564d1494b0a/kube-multus/2.log" Dec 06 05:52:14 crc kubenswrapper[4957]: I1206 05:52:14.775585 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-69r8l" event={"ID":"1384ce42-5e9a-4d6a-a299-f564d1494b0a","Type":"ContainerStarted","Data":"5d7a7e8a4a1e282c99c584d1d09917abd8b7064ad9738d2cef4d0b12267ef813"} Dec 06 05:52:15 crc kubenswrapper[4957]: I1206 05:52:15.323049 4957 scope.go:117] "RemoveContainer" containerID="1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc" Dec 06 05:52:15 crc kubenswrapper[4957]: I1206 05:52:15.794972 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-acl-logging/0.log" Dec 06 05:52:15 crc kubenswrapper[4957]: I1206 05:52:15.796413 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hlm8s_fa635ecb-5324-449e-a8f3-7a9bfdca7064/ovn-controller/0.log" Dec 06 05:52:21 crc kubenswrapper[4957]: I1206 05:52:21.132411 4957 scope.go:117] "RemoveContainer" containerID="550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.549163 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="quay.io/jetstack/cert-manager-controller:v1.14.4" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.550013 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-controller,Image:quay.io/jetstack/cert-manager-controller:v1.14.4,Command:[],Args:[--v=2 --cluster-resource-namespace=$(POD_NAMESPACE) --leader-election-namespace=kube-system --acme-http01-solver-image=quay.io/jetstack/cert-manager-acmesolver:v1.14.4 --max-concurrent-challenges=60],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:9402,Protocol:TCP,HostIP:,},ContainerPort{Name:http-healthz,HostPort:0,ContainerPort:9403,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-msgcn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{1 0 http-healthz},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:8,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000680000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-5b446d88c5-59bbt_cert-manager(ff08ac61-0abb-4f45-a647-57aca5172e36): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.551234 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-5b446d88c5-59bbt" podUID="ff08ac61-0abb-4f45-a647-57aca5172e36" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.585576 4957 scope.go:117] "RemoveContainer" containerID="526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.591966 4957 scope.go:117] "RemoveContainer" containerID="dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.613790 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="quay.io/jetstack/cert-manager-cainjector:v1.14.4" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.613973 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-cainjector,Image:quay.io/jetstack/cert-manager-cainjector:v1.14.4,Command:[],Args:[--v=2 --leader-election-namespace=kube-system],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zwf78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000680000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-cainjector-7f985d654d-6557k_cert-manager(14f39aaa-5904-4a8d-86ad-ac487456788b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.615187 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-cainjector\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" podUID="14f39aaa-5904-4a8d-86ad-ac487456788b" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.628353 4957 scope.go:117] "RemoveContainer" containerID="fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.636969 4957 scope.go:117] "RemoveContainer" containerID="1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.638716 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\": container with ID starting with 1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc not found: ID does not exist" containerID="1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.638764 4957 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\": rpc error: code = NotFound desc = could not find container \"1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc\": container with ID starting with 1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc not found: ID does not exist" containerID="1b35752b118bace8d5e3a3f1868a59460bb91e9f9366340cb315190ff68f6ffc" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.638865 4957 scope.go:117] "RemoveContainer" containerID="12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.658159 4957 scope.go:117] "RemoveContainer" containerID="12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.666482 4957 scope.go:117] "RemoveContainer" containerID="526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.666524 4957 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-ovn-metrics_ovnkube-node-hlm8s_openshift-ovn-kubernetes_fa635ecb-5324-449e-a8f3-7a9bfdca7064_0 in pod sandbox f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 from index: no such id: '12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b'" containerID="12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.666625 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b"} err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-ovn-metrics_ovnkube-node-hlm8s_openshift-ovn-kubernetes_fa635ecb-5324-449e-a8f3-7a9bfdca7064_0 in pod sandbox f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 from index: no such id: '12583cc76b9a2ac06b4faa1a0b30500b9da71e405e04a834bd0b88ca725d131b'" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.666672 4957 scope.go:117] "RemoveContainer" containerID="483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.666952 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\": container with ID starting with 526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384 not found: ID does not exist" containerID="526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.667005 4957 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\": rpc error: code = NotFound desc = could not find container \"526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384\": container with ID starting with 526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384 not found: ID does not exist" containerID="526efd739dbb98dd1adfe3c38bc6ef2bc3f9939103b50a10e168bc000342a384" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.667032 4957 scope.go:117] "RemoveContainer" containerID="483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.711253 4957 scope.go:117] "RemoveContainer" containerID="30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.711278 4957 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-node_ovnkube-node-hlm8s_openshift-ovn-kubernetes_fa635ecb-5324-449e-a8f3-7a9bfdca7064_0 in pod sandbox f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 from index: no such id: '483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d'" containerID="483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.711394 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d"} err="rpc error: code = Unknown desc = failed to delete container k8s_kube-rbac-proxy-node_ovnkube-node-hlm8s_openshift-ovn-kubernetes_fa635ecb-5324-449e-a8f3-7a9bfdca7064_0 in pod sandbox f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 from index: no such id: '483e1b313c65c29d780219f7791414cd39faa58e0462168875ea02a2befb577d'" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.711441 4957 scope.go:117] "RemoveContainer" containerID="30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.729788 4957 scope.go:117] "RemoveContainer" containerID="fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.729783 4957 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_ovn-acl-logging_ovnkube-node-hlm8s_openshift-ovn-kubernetes_fa635ecb-5324-449e-a8f3-7a9bfdca7064_0 in pod sandbox f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 from index: no such id: '30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4'" containerID="30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.729953 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4"} err="rpc error: code = Unknown desc = failed to delete container k8s_ovn-acl-logging_ovnkube-node-hlm8s_openshift-ovn-kubernetes_fa635ecb-5324-449e-a8f3-7a9bfdca7064_0 in pod sandbox f9a3717a87099e40b05c5eb7a11bf06f2d9bea5788940023108dce4fb735dcb5 from index: no such id: '30c681304ce038f548054c7f66f89b5b9c93cee444c313b1acd4d65dc7e2c9a4'" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.729993 4957 scope.go:117] "RemoveContainer" containerID="dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.730993 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\": container with ID starting with fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe not found: ID does not exist" containerID="fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.731033 4957 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\": rpc error: code = NotFound desc = could not find container \"fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe\": container with ID starting with fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe not found: ID does not exist" containerID="fd437bdc2601bb3f0d1585aaacc045de313975ca6bbab06abe4a15345cc638fe" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.731385 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\": container with ID starting with dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f not found: ID does not exist" containerID="dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.731425 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f"} err="failed to get container status \"dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\": rpc error: code = NotFound desc = could not find container \"dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f\": container with ID starting with dc5cd1052b6e8353c640227e2957392009e786a4190c1c56690fdd615b11d14f not found: ID does not exist" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.731447 4957 scope.go:117] "RemoveContainer" containerID="550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.731704 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\": container with ID starting with 550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d not found: ID does not exist" containerID="550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.731737 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d"} err="failed to get container status \"550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\": rpc error: code = NotFound desc = could not find container \"550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d\": container with ID starting with 550100a2427d58f47fc3ce6c3c01add3dc4e5c3e6d8edbacbbdb609c5e747f0d not found: ID does not exist" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.885424 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" event={"ID":"8ebad773-0957-4bb8-b583-a0cff091ff46","Type":"ContainerStarted","Data":"146538c44178fe521b5d546201a56627507d9110fcd5e2aa6794dd0d3eda1a07"} Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.885877 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.888804 4957 generic.go:334] "Generic (PLEG): container finished" podID="82e76364-26d3-4cb8-b57f-c8f03ce09a7a" containerID="9c561f771902d706d5aa20a9bc2762ce0d6e24efafd5bf0ceac840d15b009706" exitCode=0 Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.888894 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerDied","Data":"9c561f771902d706d5aa20a9bc2762ce0d6e24efafd5bf0ceac840d15b009706"} Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.888925 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"b9f759474d0b68c1763ce45088f2c91e31e9f9974a5f2c603e99ecbe674ac438"} Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.890587 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/jetstack/cert-manager-controller:v1.14.4\\\"\"" pod="cert-manager/cert-manager-5b446d88c5-59bbt" podUID="ff08ac61-0abb-4f45-a647-57aca5172e36" Dec 06 05:52:27 crc kubenswrapper[4957]: E1206 05:52:27.891187 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-cainjector\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/jetstack/cert-manager-cainjector:v1.14.4\\\"\"" pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" podUID="14f39aaa-5904-4a8d-86ad-ac487456788b" Dec 06 05:52:27 crc kubenswrapper[4957]: I1206 05:52:27.904251 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" podStartSLOduration=1.466435624 podStartE2EDuration="27.904228331s" podCreationTimestamp="2025-12-06 05:52:00 +0000 UTC" firstStartedPulling="2025-12-06 05:52:01.161571446 +0000 UTC m=+760.811839078" lastFinishedPulling="2025-12-06 05:52:27.599364153 +0000 UTC m=+787.249631785" observedRunningTime="2025-12-06 05:52:27.90015577 +0000 UTC m=+787.550423422" watchObservedRunningTime="2025-12-06 05:52:27.904228331 +0000 UTC m=+787.554495963" Dec 06 05:52:28 crc kubenswrapper[4957]: I1206 05:52:28.898206 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"7187a459aba436bb5e993f1388b59ed1213030c8c71827cbacded2e2d8ffffa3"} Dec 06 05:52:29 crc kubenswrapper[4957]: I1206 05:52:29.910584 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"1cc1eb0135d5ddd1f01be76029d1338ee2aa9416d69298907f503990f562c1dd"} Dec 06 05:52:29 crc kubenswrapper[4957]: I1206 05:52:29.910657 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"d3d7a641b825c84becae4f506fe2853d58bbd8b932bf99438cd19f7944051333"} Dec 06 05:52:30 crc kubenswrapper[4957]: I1206 05:52:30.917622 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"0f0106f3de451355c02079258e4e7468a94136c2ae3ce5cd4130dfbd12905143"} Dec 06 05:52:32 crc kubenswrapper[4957]: I1206 05:52:32.934883 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"39bb265a7a7bcb300d5d65133832fa9757dedae871d94858c5bee0df18e0917f"} Dec 06 05:52:33 crc kubenswrapper[4957]: I1206 05:52:33.944729 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"9b8901d1255799df404601dec836583bcddf6c58c1dcf4b34eee959053c125cd"} Dec 06 05:52:35 crc kubenswrapper[4957]: I1206 05:52:35.879627 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-b256m" Dec 06 05:52:35 crc kubenswrapper[4957]: I1206 05:52:35.962071 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"3b04e51373a30b944b9e92f04ca239dd3badca637477d6c10a819057807fcc7b"} Dec 06 05:52:37 crc kubenswrapper[4957]: I1206 05:52:37.978450 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" event={"ID":"82e76364-26d3-4cb8-b57f-c8f03ce09a7a","Type":"ContainerStarted","Data":"3b0c9951759c1fca03e7f5a7cb8ecc5d681177fd097b9bb4e75872522b7d3e09"} Dec 06 05:52:37 crc kubenswrapper[4957]: I1206 05:52:37.978844 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:37 crc kubenswrapper[4957]: I1206 05:52:37.978863 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:37 crc kubenswrapper[4957]: I1206 05:52:37.979076 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:38 crc kubenswrapper[4957]: I1206 05:52:38.012881 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" podStartSLOduration=25.012859594 podStartE2EDuration="25.012859594s" podCreationTimestamp="2025-12-06 05:52:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:52:38.007881187 +0000 UTC m=+797.658148839" watchObservedRunningTime="2025-12-06 05:52:38.012859594 +0000 UTC m=+797.663127236" Dec 06 05:52:38 crc kubenswrapper[4957]: I1206 05:52:38.130145 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:38 crc kubenswrapper[4957]: I1206 05:52:38.139546 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:52:40 crc kubenswrapper[4957]: I1206 05:52:40.180573 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:52:40 crc kubenswrapper[4957]: I1206 05:52:40.181020 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:52:40 crc kubenswrapper[4957]: I1206 05:52:40.181082 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:52:40 crc kubenswrapper[4957]: I1206 05:52:40.181719 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b0c4f58bdcf5290181ec1a457cbeb02772611a1bd68d33a39e309f1aaee0927"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:52:40 crc kubenswrapper[4957]: I1206 05:52:40.181785 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://0b0c4f58bdcf5290181ec1a457cbeb02772611a1bd68d33a39e309f1aaee0927" gracePeriod=600 Dec 06 05:52:42 crc kubenswrapper[4957]: I1206 05:52:42.007410 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="0b0c4f58bdcf5290181ec1a457cbeb02772611a1bd68d33a39e309f1aaee0927" exitCode=0 Dec 06 05:52:42 crc kubenswrapper[4957]: I1206 05:52:42.007510 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"0b0c4f58bdcf5290181ec1a457cbeb02772611a1bd68d33a39e309f1aaee0927"} Dec 06 05:52:42 crc kubenswrapper[4957]: I1206 05:52:42.007942 4957 scope.go:117] "RemoveContainer" containerID="af5d9f710568e718f75a967bf395611eae1b65dcacd5acd1a647b8480dd1df56" Dec 06 05:52:43 crc kubenswrapper[4957]: I1206 05:52:43.587511 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" podUID="82e76364-26d3-4cb8-b57f-c8f03ce09a7a" containerName="ovnkube-controller" probeResult="failure" output="" Dec 06 05:52:44 crc kubenswrapper[4957]: I1206 05:52:44.020482 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" event={"ID":"14f39aaa-5904-4a8d-86ad-ac487456788b","Type":"ContainerStarted","Data":"d689edb4e26780e10a18507e293ee3ab0536e59ac34a18832c37ab1d8b705970"} Dec 06 05:52:44 crc kubenswrapper[4957]: I1206 05:52:44.022093 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-59bbt" event={"ID":"ff08ac61-0abb-4f45-a647-57aca5172e36","Type":"ContainerStarted","Data":"394db316f90385e3bbcc44056c87b3235d5552abbc06487913d6e0002de60f5e"} Dec 06 05:52:44 crc kubenswrapper[4957]: I1206 05:52:44.024708 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"c315dd48c0a384ca90d94b9ce0d09121f3d64da936f2696223ce8145d1b5a516"} Dec 06 05:52:44 crc kubenswrapper[4957]: I1206 05:52:44.035775 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-6557k" podStartSLOduration=2.916055727 podStartE2EDuration="44.035756102s" podCreationTimestamp="2025-12-06 05:52:00 +0000 UTC" firstStartedPulling="2025-12-06 05:52:01.125022628 +0000 UTC m=+760.775290260" lastFinishedPulling="2025-12-06 05:52:42.244722973 +0000 UTC m=+801.894990635" observedRunningTime="2025-12-06 05:52:44.034080956 +0000 UTC m=+803.684348588" watchObservedRunningTime="2025-12-06 05:52:44.035756102 +0000 UTC m=+803.686023734" Dec 06 05:52:44 crc kubenswrapper[4957]: I1206 05:52:44.067410 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-59bbt" podStartSLOduration=1.625264042 podStartE2EDuration="44.067392594s" podCreationTimestamp="2025-12-06 05:52:00 +0000 UTC" firstStartedPulling="2025-12-06 05:52:01.196628684 +0000 UTC m=+760.846896316" lastFinishedPulling="2025-12-06 05:52:43.638757236 +0000 UTC m=+803.289024868" observedRunningTime="2025-12-06 05:52:44.064978487 +0000 UTC m=+803.715246129" watchObservedRunningTime="2025-12-06 05:52:44.067392594 +0000 UTC m=+803.717660226" Dec 06 05:53:13 crc kubenswrapper[4957]: I1206 05:53:13.594365 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nwk7m" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.497303 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn"] Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.500021 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.506143 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.506997 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn"] Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.524419 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.524561 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvvmz\" (UniqueName: \"kubernetes.io/projected/0c166589-a548-437d-bdeb-b4cefa7fa6bc-kube-api-access-xvvmz\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.524683 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.626524 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.626669 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvvmz\" (UniqueName: \"kubernetes.io/projected/0c166589-a548-437d-bdeb-b4cefa7fa6bc-kube-api-access-xvvmz\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.626799 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.627508 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.628140 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.653951 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvvmz\" (UniqueName: \"kubernetes.io/projected/0c166589-a548-437d-bdeb-b4cefa7fa6bc-kube-api-access-xvvmz\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:32 crc kubenswrapper[4957]: I1206 05:53:32.821276 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:53:33 crc kubenswrapper[4957]: I1206 05:53:33.120202 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn"] Dec 06 05:53:33 crc kubenswrapper[4957]: I1206 05:53:33.371361 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerStarted","Data":"3f0e6044df49763cb4c4c920de2c46257390a8246dda50570bd39108838814a9"} Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.378736 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerStarted","Data":"3d1af040c670129c4f8ee7fc99c3fc5569baf7931b9afcd0b503e5ac4b5cb5ac"} Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.700008 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-njsbw"] Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.703623 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-njsbw"] Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.703887 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.816653 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-utilities\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.816707 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqp2h\" (UniqueName: \"kubernetes.io/projected/0e1b36d1-ed58-439e-ad42-fe17649b18d6-kube-api-access-vqp2h\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.816811 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-catalog-content\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.918608 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-catalog-content\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.918721 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-utilities\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.918767 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqp2h\" (UniqueName: \"kubernetes.io/projected/0e1b36d1-ed58-439e-ad42-fe17649b18d6-kube-api-access-vqp2h\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.919355 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-catalog-content\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.920018 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-utilities\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:34 crc kubenswrapper[4957]: I1206 05:53:34.952182 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqp2h\" (UniqueName: \"kubernetes.io/projected/0e1b36d1-ed58-439e-ad42-fe17649b18d6-kube-api-access-vqp2h\") pod \"redhat-operators-njsbw\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:35 crc kubenswrapper[4957]: I1206 05:53:35.032240 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:53:35 crc kubenswrapper[4957]: I1206 05:53:35.385628 4957 generic.go:334] "Generic (PLEG): container finished" podID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerID="3d1af040c670129c4f8ee7fc99c3fc5569baf7931b9afcd0b503e5ac4b5cb5ac" exitCode=0 Dec 06 05:53:35 crc kubenswrapper[4957]: I1206 05:53:35.385677 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerDied","Data":"3d1af040c670129c4f8ee7fc99c3fc5569baf7931b9afcd0b503e5ac4b5cb5ac"} Dec 06 05:53:35 crc kubenswrapper[4957]: I1206 05:53:35.458235 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-njsbw"] Dec 06 05:53:35 crc kubenswrapper[4957]: W1206 05:53:35.461270 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e1b36d1_ed58_439e_ad42_fe17649b18d6.slice/crio-2d04a8d08d57238e2433720a0ea1ac15ba5b4fcccdff7f8465ad7a539a354381 WatchSource:0}: Error finding container 2d04a8d08d57238e2433720a0ea1ac15ba5b4fcccdff7f8465ad7a539a354381: Status 404 returned error can't find the container with id 2d04a8d08d57238e2433720a0ea1ac15ba5b4fcccdff7f8465ad7a539a354381 Dec 06 05:53:36 crc kubenswrapper[4957]: I1206 05:53:36.399152 4957 generic.go:334] "Generic (PLEG): container finished" podID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerID="83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b" exitCode=0 Dec 06 05:53:36 crc kubenswrapper[4957]: I1206 05:53:36.399353 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njsbw" event={"ID":"0e1b36d1-ed58-439e-ad42-fe17649b18d6","Type":"ContainerDied","Data":"83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b"} Dec 06 05:53:36 crc kubenswrapper[4957]: I1206 05:53:36.399743 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njsbw" event={"ID":"0e1b36d1-ed58-439e-ad42-fe17649b18d6","Type":"ContainerStarted","Data":"2d04a8d08d57238e2433720a0ea1ac15ba5b4fcccdff7f8465ad7a539a354381"} Dec 06 05:53:41 crc kubenswrapper[4957]: I1206 05:53:41.434206 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njsbw" event={"ID":"0e1b36d1-ed58-439e-ad42-fe17649b18d6","Type":"ContainerStarted","Data":"43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7"} Dec 06 05:53:43 crc kubenswrapper[4957]: I1206 05:53:43.447677 4957 generic.go:334] "Generic (PLEG): container finished" podID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerID="43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7" exitCode=0 Dec 06 05:53:43 crc kubenswrapper[4957]: I1206 05:53:43.447780 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njsbw" event={"ID":"0e1b36d1-ed58-439e-ad42-fe17649b18d6","Type":"ContainerDied","Data":"43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7"} Dec 06 05:53:45 crc kubenswrapper[4957]: I1206 05:53:45.460997 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerStarted","Data":"1cc6093c1d66b4b19b2e95c127488546639cb6faae3d076a2d0e95c6d32a45d4"} Dec 06 05:53:46 crc kubenswrapper[4957]: I1206 05:53:46.471599 4957 generic.go:334] "Generic (PLEG): container finished" podID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerID="1cc6093c1d66b4b19b2e95c127488546639cb6faae3d076a2d0e95c6d32a45d4" exitCode=0 Dec 06 05:53:46 crc kubenswrapper[4957]: I1206 05:53:46.471672 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerDied","Data":"1cc6093c1d66b4b19b2e95c127488546639cb6faae3d076a2d0e95c6d32a45d4"} Dec 06 05:53:47 crc kubenswrapper[4957]: I1206 05:53:47.489428 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerStarted","Data":"fe52dddde7bfd189f62600d8f4e791f2ce28adae494deccc7714280bde8c7ab4"} Dec 06 05:53:48 crc kubenswrapper[4957]: I1206 05:53:48.514269 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" podStartSLOduration=6.880789732 podStartE2EDuration="16.51425162s" podCreationTimestamp="2025-12-06 05:53:32 +0000 UTC" firstStartedPulling="2025-12-06 05:53:35.38691202 +0000 UTC m=+855.037179652" lastFinishedPulling="2025-12-06 05:53:45.020373898 +0000 UTC m=+864.670641540" observedRunningTime="2025-12-06 05:53:48.512374319 +0000 UTC m=+868.162641991" watchObservedRunningTime="2025-12-06 05:53:48.51425162 +0000 UTC m=+868.164519262" Dec 06 05:53:57 crc kubenswrapper[4957]: I1206 05:53:57.967164 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vsrb4"] Dec 06 05:53:57 crc kubenswrapper[4957]: I1206 05:53:57.976990 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:57 crc kubenswrapper[4957]: I1206 05:53:57.987233 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vsrb4"] Dec 06 05:53:57 crc kubenswrapper[4957]: I1206 05:53:57.995853 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-catalog-content\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:57 crc kubenswrapper[4957]: I1206 05:53:57.995913 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-utilities\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:57 crc kubenswrapper[4957]: I1206 05:53:57.995937 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jkc8\" (UniqueName: \"kubernetes.io/projected/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-kube-api-access-4jkc8\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:58 crc kubenswrapper[4957]: I1206 05:53:58.097700 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-catalog-content\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:58 crc kubenswrapper[4957]: I1206 05:53:58.098124 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-utilities\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:58 crc kubenswrapper[4957]: I1206 05:53:58.098159 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jkc8\" (UniqueName: \"kubernetes.io/projected/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-kube-api-access-4jkc8\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:58 crc kubenswrapper[4957]: I1206 05:53:58.098284 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-catalog-content\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:58 crc kubenswrapper[4957]: I1206 05:53:58.098696 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-utilities\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:58 crc kubenswrapper[4957]: I1206 05:53:58.131120 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jkc8\" (UniqueName: \"kubernetes.io/projected/25410e1a-80c2-4f9f-9c62-2bd939fa21c2-kube-api-access-4jkc8\") pod \"community-operators-vsrb4\" (UID: \"25410e1a-80c2-4f9f-9c62-2bd939fa21c2\") " pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:53:58 crc kubenswrapper[4957]: I1206 05:53:58.296905 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:54:01 crc kubenswrapper[4957]: I1206 05:54:01.587405 4957 generic.go:334] "Generic (PLEG): container finished" podID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerID="fe52dddde7bfd189f62600d8f4e791f2ce28adae494deccc7714280bde8c7ab4" exitCode=0 Dec 06 05:54:01 crc kubenswrapper[4957]: I1206 05:54:01.587463 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerDied","Data":"fe52dddde7bfd189f62600d8f4e791f2ce28adae494deccc7714280bde8c7ab4"} Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:02.913643 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.002152 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvvmz\" (UniqueName: \"kubernetes.io/projected/0c166589-a548-437d-bdeb-b4cefa7fa6bc-kube-api-access-xvvmz\") pod \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.002233 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-util\") pod \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.002277 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-bundle\") pod \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\" (UID: \"0c166589-a548-437d-bdeb-b4cefa7fa6bc\") " Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.003595 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-bundle" (OuterVolumeSpecName: "bundle") pod "0c166589-a548-437d-bdeb-b4cefa7fa6bc" (UID: "0c166589-a548-437d-bdeb-b4cefa7fa6bc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.011457 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c166589-a548-437d-bdeb-b4cefa7fa6bc-kube-api-access-xvvmz" (OuterVolumeSpecName: "kube-api-access-xvvmz") pod "0c166589-a548-437d-bdeb-b4cefa7fa6bc" (UID: "0c166589-a548-437d-bdeb-b4cefa7fa6bc"). InnerVolumeSpecName "kube-api-access-xvvmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.025674 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-util" (OuterVolumeSpecName: "util") pod "0c166589-a548-437d-bdeb-b4cefa7fa6bc" (UID: "0c166589-a548-437d-bdeb-b4cefa7fa6bc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.104953 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvvmz\" (UniqueName: \"kubernetes.io/projected/0c166589-a548-437d-bdeb-b4cefa7fa6bc-kube-api-access-xvvmz\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.105001 4957 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-util\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.105020 4957 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c166589-a548-437d-bdeb-b4cefa7fa6bc-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.611603 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" event={"ID":"0c166589-a548-437d-bdeb-b4cefa7fa6bc","Type":"ContainerDied","Data":"3f0e6044df49763cb4c4c920de2c46257390a8246dda50570bd39108838814a9"} Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.612170 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f0e6044df49763cb4c4c920de2c46257390a8246dda50570bd39108838814a9" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:03.611746 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn" Dec 06 05:54:05 crc kubenswrapper[4957]: I1206 05:54:05.936018 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vsrb4"] Dec 06 05:54:06 crc kubenswrapper[4957]: I1206 05:54:06.633697 4957 generic.go:334] "Generic (PLEG): container finished" podID="25410e1a-80c2-4f9f-9c62-2bd939fa21c2" containerID="a813c7b0698dc47cd1ef8cd82728ad89872eea8951ed6b1682308ae2ad282729" exitCode=0 Dec 06 05:54:06 crc kubenswrapper[4957]: I1206 05:54:06.633801 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsrb4" event={"ID":"25410e1a-80c2-4f9f-9c62-2bd939fa21c2","Type":"ContainerDied","Data":"a813c7b0698dc47cd1ef8cd82728ad89872eea8951ed6b1682308ae2ad282729"} Dec 06 05:54:06 crc kubenswrapper[4957]: I1206 05:54:06.634015 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsrb4" event={"ID":"25410e1a-80c2-4f9f-9c62-2bd939fa21c2","Type":"ContainerStarted","Data":"58de08527b2352d46efbddfc73d8d0a2bbd6f7ad3a0a2d27b0f7a034e11dc1ac"} Dec 06 05:54:06 crc kubenswrapper[4957]: I1206 05:54:06.638377 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njsbw" event={"ID":"0e1b36d1-ed58-439e-ad42-fe17649b18d6","Type":"ContainerStarted","Data":"bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832"} Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.503934 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-njsbw" podStartSLOduration=5.160106525 podStartE2EDuration="34.503911363s" podCreationTimestamp="2025-12-06 05:53:34 +0000 UTC" firstStartedPulling="2025-12-06 05:53:36.400789715 +0000 UTC m=+856.051057347" lastFinishedPulling="2025-12-06 05:54:05.744594553 +0000 UTC m=+885.394862185" observedRunningTime="2025-12-06 05:54:06.679147923 +0000 UTC m=+886.329415575" watchObservedRunningTime="2025-12-06 05:54:08.503911363 +0000 UTC m=+888.154179015" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.507667 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5"] Dec 06 05:54:08 crc kubenswrapper[4957]: E1206 05:54:08.507952 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerName="util" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.507976 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerName="util" Dec 06 05:54:08 crc kubenswrapper[4957]: E1206 05:54:08.507996 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerName="pull" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.508004 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerName="pull" Dec 06 05:54:08 crc kubenswrapper[4957]: E1206 05:54:08.508016 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerName="extract" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.508023 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerName="extract" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.508155 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c166589-a548-437d-bdeb-b4cefa7fa6bc" containerName="extract" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.508684 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.510721 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.510944 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mwgqz" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.511103 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.525569 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5"] Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.583348 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5kwg\" (UniqueName: \"kubernetes.io/projected/44a5fc06-b6e1-4852-a528-f92117991c7a-kube-api-access-t5kwg\") pod \"nmstate-operator-5b5b58f5c8-2d6j5\" (UID: \"44a5fc06-b6e1-4852-a528-f92117991c7a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.684550 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5kwg\" (UniqueName: \"kubernetes.io/projected/44a5fc06-b6e1-4852-a528-f92117991c7a-kube-api-access-t5kwg\") pod \"nmstate-operator-5b5b58f5c8-2d6j5\" (UID: \"44a5fc06-b6e1-4852-a528-f92117991c7a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.716281 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5kwg\" (UniqueName: \"kubernetes.io/projected/44a5fc06-b6e1-4852-a528-f92117991c7a-kube-api-access-t5kwg\") pod \"nmstate-operator-5b5b58f5c8-2d6j5\" (UID: \"44a5fc06-b6e1-4852-a528-f92117991c7a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" Dec 06 05:54:08 crc kubenswrapper[4957]: I1206 05:54:08.825210 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" Dec 06 05:54:09 crc kubenswrapper[4957]: I1206 05:54:09.118018 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5"] Dec 06 05:54:09 crc kubenswrapper[4957]: I1206 05:54:09.660054 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" event={"ID":"44a5fc06-b6e1-4852-a528-f92117991c7a","Type":"ContainerStarted","Data":"3ebdd814f32c20f0ad08f93251af16c2b5cd4cec9bb6ffa972e74313a75878bf"} Dec 06 05:54:15 crc kubenswrapper[4957]: I1206 05:54:15.033088 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:54:15 crc kubenswrapper[4957]: I1206 05:54:15.033780 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:54:15 crc kubenswrapper[4957]: I1206 05:54:15.196634 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:54:15 crc kubenswrapper[4957]: I1206 05:54:15.725269 4957 generic.go:334] "Generic (PLEG): container finished" podID="25410e1a-80c2-4f9f-9c62-2bd939fa21c2" containerID="a8a30684627b6d043a6398a0d4437b406fa60c58bd550558b2e0da8efe772120" exitCode=0 Dec 06 05:54:15 crc kubenswrapper[4957]: I1206 05:54:15.725367 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsrb4" event={"ID":"25410e1a-80c2-4f9f-9c62-2bd939fa21c2","Type":"ContainerDied","Data":"a8a30684627b6d043a6398a0d4437b406fa60c58bd550558b2e0da8efe772120"} Dec 06 05:54:15 crc kubenswrapper[4957]: I1206 05:54:15.771068 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.155000 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zznbp"] Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.156119 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.170613 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zznbp"] Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.221601 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-utilities\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.221679 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-catalog-content\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.221713 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcl6r\" (UniqueName: \"kubernetes.io/projected/2c524bd5-ac23-4921-a527-a6e644e3aa09-kube-api-access-xcl6r\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.335883 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-utilities\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.335934 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-catalog-content\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.335958 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcl6r\" (UniqueName: \"kubernetes.io/projected/2c524bd5-ac23-4921-a527-a6e644e3aa09-kube-api-access-xcl6r\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.336822 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-utilities\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.336947 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-catalog-content\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.374003 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcl6r\" (UniqueName: \"kubernetes.io/projected/2c524bd5-ac23-4921-a527-a6e644e3aa09-kube-api-access-xcl6r\") pod \"redhat-marketplace-zznbp\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.473207 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.671331 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zznbp"] Dec 06 05:54:17 crc kubenswrapper[4957]: W1206 05:54:17.681154 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c524bd5_ac23_4921_a527_a6e644e3aa09.slice/crio-1a5fa559f50888b573ae91a97ec67509e6fb8c461fcf45067fb829eda17067ab WatchSource:0}: Error finding container 1a5fa559f50888b573ae91a97ec67509e6fb8c461fcf45067fb829eda17067ab: Status 404 returned error can't find the container with id 1a5fa559f50888b573ae91a97ec67509e6fb8c461fcf45067fb829eda17067ab Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.738817 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zznbp" event={"ID":"2c524bd5-ac23-4921-a527-a6e644e3aa09","Type":"ContainerStarted","Data":"1a5fa559f50888b573ae91a97ec67509e6fb8c461fcf45067fb829eda17067ab"} Dec 06 05:54:17 crc kubenswrapper[4957]: I1206 05:54:17.740849 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" event={"ID":"44a5fc06-b6e1-4852-a528-f92117991c7a","Type":"ContainerStarted","Data":"7a423e28f4354e41ab6f2b19c8edba6c3e92568e1d6267320157a6a40afd372b"} Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.147985 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-njsbw"] Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.148626 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-njsbw" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="registry-server" containerID="cri-o://bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832" gracePeriod=2 Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.492464 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.656700 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-utilities\") pod \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.656761 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqp2h\" (UniqueName: \"kubernetes.io/projected/0e1b36d1-ed58-439e-ad42-fe17649b18d6-kube-api-access-vqp2h\") pod \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.656918 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-catalog-content\") pod \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\" (UID: \"0e1b36d1-ed58-439e-ad42-fe17649b18d6\") " Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.658241 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-utilities" (OuterVolumeSpecName: "utilities") pod "0e1b36d1-ed58-439e-ad42-fe17649b18d6" (UID: "0e1b36d1-ed58-439e-ad42-fe17649b18d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.666760 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e1b36d1-ed58-439e-ad42-fe17649b18d6-kube-api-access-vqp2h" (OuterVolumeSpecName: "kube-api-access-vqp2h") pod "0e1b36d1-ed58-439e-ad42-fe17649b18d6" (UID: "0e1b36d1-ed58-439e-ad42-fe17649b18d6"). InnerVolumeSpecName "kube-api-access-vqp2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.750705 4957 generic.go:334] "Generic (PLEG): container finished" podID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerID="bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832" exitCode=0 Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.750781 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-njsbw" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.750814 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njsbw" event={"ID":"0e1b36d1-ed58-439e-ad42-fe17649b18d6","Type":"ContainerDied","Data":"bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832"} Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.751324 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-njsbw" event={"ID":"0e1b36d1-ed58-439e-ad42-fe17649b18d6","Type":"ContainerDied","Data":"2d04a8d08d57238e2433720a0ea1ac15ba5b4fcccdff7f8465ad7a539a354381"} Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.751374 4957 scope.go:117] "RemoveContainer" containerID="bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.754194 4957 generic.go:334] "Generic (PLEG): container finished" podID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerID="f90dba1cdd20d8b0584e55c9a5f99f9299893fcdb710762e1819b4761ee16ba9" exitCode=0 Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.754225 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zznbp" event={"ID":"2c524bd5-ac23-4921-a527-a6e644e3aa09","Type":"ContainerDied","Data":"f90dba1cdd20d8b0584e55c9a5f99f9299893fcdb710762e1819b4761ee16ba9"} Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.759156 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e1b36d1-ed58-439e-ad42-fe17649b18d6" (UID: "0e1b36d1-ed58-439e-ad42-fe17649b18d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.761004 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.761411 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b36d1-ed58-439e-ad42-fe17649b18d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.761539 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqp2h\" (UniqueName: \"kubernetes.io/projected/0e1b36d1-ed58-439e-ad42-fe17649b18d6-kube-api-access-vqp2h\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.761786 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsrb4" event={"ID":"25410e1a-80c2-4f9f-9c62-2bd939fa21c2","Type":"ContainerStarted","Data":"13651b465289ab393944bc50ffa5f89bbf5bd7856ff851b770deac2fbdce0ef8"} Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.776383 4957 scope.go:117] "RemoveContainer" containerID="43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.802497 4957 scope.go:117] "RemoveContainer" containerID="83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.803604 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2d6j5" podStartSLOduration=4.052804902 podStartE2EDuration="10.80358265s" podCreationTimestamp="2025-12-06 05:54:08 +0000 UTC" firstStartedPulling="2025-12-06 05:54:09.13189657 +0000 UTC m=+888.782164202" lastFinishedPulling="2025-12-06 05:54:15.882674318 +0000 UTC m=+895.532941950" observedRunningTime="2025-12-06 05:54:18.790945852 +0000 UTC m=+898.441213504" watchObservedRunningTime="2025-12-06 05:54:18.80358265 +0000 UTC m=+898.453850272" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.822458 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vsrb4" podStartSLOduration=10.360460388 podStartE2EDuration="21.822436819s" podCreationTimestamp="2025-12-06 05:53:57 +0000 UTC" firstStartedPulling="2025-12-06 05:54:06.637006862 +0000 UTC m=+886.287274494" lastFinishedPulling="2025-12-06 05:54:18.098983293 +0000 UTC m=+897.749250925" observedRunningTime="2025-12-06 05:54:18.818862501 +0000 UTC m=+898.469130163" watchObservedRunningTime="2025-12-06 05:54:18.822436819 +0000 UTC m=+898.472704451" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.831953 4957 scope.go:117] "RemoveContainer" containerID="bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832" Dec 06 05:54:18 crc kubenswrapper[4957]: E1206 05:54:18.840014 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832\": container with ID starting with bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832 not found: ID does not exist" containerID="bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.840082 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832"} err="failed to get container status \"bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832\": rpc error: code = NotFound desc = could not find container \"bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832\": container with ID starting with bb1bedd5e857b27e44e73e60fd8580e45ee9ade6f347a3c8d61d749823d54832 not found: ID does not exist" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.840114 4957 scope.go:117] "RemoveContainer" containerID="43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7" Dec 06 05:54:18 crc kubenswrapper[4957]: E1206 05:54:18.841877 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7\": container with ID starting with 43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7 not found: ID does not exist" containerID="43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.841904 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7"} err="failed to get container status \"43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7\": rpc error: code = NotFound desc = could not find container \"43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7\": container with ID starting with 43821a4ff71fe16b95b79d4b8cc17b7f7fd40cd5f79b38098cd6b0de3f34ecd7 not found: ID does not exist" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.841921 4957 scope.go:117] "RemoveContainer" containerID="83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b" Dec 06 05:54:18 crc kubenswrapper[4957]: E1206 05:54:18.842539 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b\": container with ID starting with 83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b not found: ID does not exist" containerID="83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b" Dec 06 05:54:18 crc kubenswrapper[4957]: I1206 05:54:18.842566 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b"} err="failed to get container status \"83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b\": rpc error: code = NotFound desc = could not find container \"83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b\": container with ID starting with 83acf9ce293f86796cfe496904d84939dd3b42b80dc60752b70ad3ebd7f50d6b not found: ID does not exist" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.077860 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-njsbw"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.082135 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-njsbw"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.767968 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd"] Dec 06 05:54:19 crc kubenswrapper[4957]: E1206 05:54:19.768476 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="extract-content" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.768489 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="extract-content" Dec 06 05:54:19 crc kubenswrapper[4957]: E1206 05:54:19.768499 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="extract-utilities" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.768505 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="extract-utilities" Dec 06 05:54:19 crc kubenswrapper[4957]: E1206 05:54:19.768519 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="registry-server" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.768525 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="registry-server" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.768617 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" containerName="registry-server" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.769000 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.771449 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.771538 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-9vvs9" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.772746 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.772998 4957 generic.go:334] "Generic (PLEG): container finished" podID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerID="2b8fbe80034c7602898fd1afffffe82d18ad3adc9f262feec00ec340679fd3f3" exitCode=0 Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.773992 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zznbp" event={"ID":"2c524bd5-ac23-4921-a527-a6e644e3aa09","Type":"ContainerDied","Data":"2b8fbe80034c7602898fd1afffffe82d18ad3adc9f262feec00ec340679fd3f3"} Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.774100 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.790382 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.800131 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-x8lpv"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.800921 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.805964 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.876268 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9shhj\" (UniqueName: \"kubernetes.io/projected/a60b6144-6a1f-4a86-ae5e-84d5771a8c9c-kube-api-access-9shhj\") pod \"nmstate-metrics-7f946cbc9-mlsvn\" (UID: \"a60b6144-6a1f-4a86-ae5e-84d5771a8c9c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.876339 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j5pk\" (UniqueName: \"kubernetes.io/projected/f85d16c2-33be-4772-a597-fa4393b5846c-kube-api-access-9j5pk\") pod \"nmstate-webhook-5f6d4c5ccb-6r8nd\" (UID: \"f85d16c2-33be-4772-a597-fa4393b5846c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.876367 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f85d16c2-33be-4772-a597-fa4393b5846c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6r8nd\" (UID: \"f85d16c2-33be-4772-a597-fa4393b5846c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.876460 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-dbus-socket\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.876484 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-nmstate-lock\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.876510 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttgxf\" (UniqueName: \"kubernetes.io/projected/211db7e0-522a-4635-a279-1d58502c1d1f-kube-api-access-ttgxf\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.876549 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-ovs-socket\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.894644 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.895327 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.897630 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.898032 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.899931 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-xh9pd" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.925106 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f"] Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978257 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j5pk\" (UniqueName: \"kubernetes.io/projected/f85d16c2-33be-4772-a597-fa4393b5846c-kube-api-access-9j5pk\") pod \"nmstate-webhook-5f6d4c5ccb-6r8nd\" (UID: \"f85d16c2-33be-4772-a597-fa4393b5846c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978311 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978338 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f85d16c2-33be-4772-a597-fa4393b5846c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6r8nd\" (UID: \"f85d16c2-33be-4772-a597-fa4393b5846c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978382 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-dbus-socket\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978401 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978419 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-nmstate-lock\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978437 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttgxf\" (UniqueName: \"kubernetes.io/projected/211db7e0-522a-4635-a279-1d58502c1d1f-kube-api-access-ttgxf\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978455 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzxjx\" (UniqueName: \"kubernetes.io/projected/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-kube-api-access-xzxjx\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978476 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-ovs-socket\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978516 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9shhj\" (UniqueName: \"kubernetes.io/projected/a60b6144-6a1f-4a86-ae5e-84d5771a8c9c-kube-api-access-9shhj\") pod \"nmstate-metrics-7f946cbc9-mlsvn\" (UID: \"a60b6144-6a1f-4a86-ae5e-84d5771a8c9c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" Dec 06 05:54:19 crc kubenswrapper[4957]: E1206 05:54:19.978558 4957 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978615 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-ovs-socket\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978581 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-nmstate-lock\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:19 crc kubenswrapper[4957]: E1206 05:54:19.978647 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f85d16c2-33be-4772-a597-fa4393b5846c-tls-key-pair podName:f85d16c2-33be-4772-a597-fa4393b5846c nodeName:}" failed. No retries permitted until 2025-12-06 05:54:20.478624674 +0000 UTC m=+900.128892376 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f85d16c2-33be-4772-a597-fa4393b5846c-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-6r8nd" (UID: "f85d16c2-33be-4772-a597-fa4393b5846c") : secret "openshift-nmstate-webhook" not found Dec 06 05:54:19 crc kubenswrapper[4957]: I1206 05:54:19.978691 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/211db7e0-522a-4635-a279-1d58502c1d1f-dbus-socket\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.000237 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttgxf\" (UniqueName: \"kubernetes.io/projected/211db7e0-522a-4635-a279-1d58502c1d1f-kube-api-access-ttgxf\") pod \"nmstate-handler-x8lpv\" (UID: \"211db7e0-522a-4635-a279-1d58502c1d1f\") " pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.000260 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j5pk\" (UniqueName: \"kubernetes.io/projected/f85d16c2-33be-4772-a597-fa4393b5846c-kube-api-access-9j5pk\") pod \"nmstate-webhook-5f6d4c5ccb-6r8nd\" (UID: \"f85d16c2-33be-4772-a597-fa4393b5846c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.002124 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9shhj\" (UniqueName: \"kubernetes.io/projected/a60b6144-6a1f-4a86-ae5e-84d5771a8c9c-kube-api-access-9shhj\") pod \"nmstate-metrics-7f946cbc9-mlsvn\" (UID: \"a60b6144-6a1f-4a86-ae5e-84d5771a8c9c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.079867 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.079979 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.080029 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzxjx\" (UniqueName: \"kubernetes.io/projected/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-kube-api-access-xzxjx\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: E1206 05:54:20.080472 4957 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 06 05:54:20 crc kubenswrapper[4957]: E1206 05:54:20.080568 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-plugin-serving-cert podName:cbf7952f-0cb5-4873-a116-b878ccdbf7e2 nodeName:}" failed. No retries permitted until 2025-12-06 05:54:20.580545631 +0000 UTC m=+900.230813263 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-wtc5f" (UID: "cbf7952f-0cb5-4873-a116-b878ccdbf7e2") : secret "plugin-serving-cert" not found Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.080995 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.096349 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.108785 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzxjx\" (UniqueName: \"kubernetes.io/projected/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-kube-api-access-xzxjx\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.119714 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.195991 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6f54f5f4f4-xnz7z"] Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.197025 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.264575 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f54f5f4f4-xnz7z"] Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.286450 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv67p\" (UniqueName: \"kubernetes.io/projected/5ee5dadd-4803-4a63-8d41-651780b2d689-kube-api-access-fv67p\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.286509 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-oauth-serving-cert\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.286623 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5ee5dadd-4803-4a63-8d41-651780b2d689-console-oauth-config\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.286662 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-service-ca\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.286697 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-console-config\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.286721 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ee5dadd-4803-4a63-8d41-651780b2d689-console-serving-cert\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.286784 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-trusted-ca-bundle\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.383675 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn"] Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.388003 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-trusted-ca-bundle\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.388074 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv67p\" (UniqueName: \"kubernetes.io/projected/5ee5dadd-4803-4a63-8d41-651780b2d689-kube-api-access-fv67p\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.388103 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-oauth-serving-cert\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.388134 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5ee5dadd-4803-4a63-8d41-651780b2d689-console-oauth-config\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.388168 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-service-ca\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.388184 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-console-config\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.388206 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ee5dadd-4803-4a63-8d41-651780b2d689-console-serving-cert\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.392214 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-service-ca\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.392857 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5ee5dadd-4803-4a63-8d41-651780b2d689-console-oauth-config\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.393052 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-console-config\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.393358 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-oauth-serving-cert\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.393681 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ee5dadd-4803-4a63-8d41-651780b2d689-trusted-ca-bundle\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.394175 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ee5dadd-4803-4a63-8d41-651780b2d689-console-serving-cert\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.409758 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv67p\" (UniqueName: \"kubernetes.io/projected/5ee5dadd-4803-4a63-8d41-651780b2d689-kube-api-access-fv67p\") pod \"console-6f54f5f4f4-xnz7z\" (UID: \"5ee5dadd-4803-4a63-8d41-651780b2d689\") " pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.489542 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f85d16c2-33be-4772-a597-fa4393b5846c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6r8nd\" (UID: \"f85d16c2-33be-4772-a597-fa4393b5846c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.492941 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f85d16c2-33be-4772-a597-fa4393b5846c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-6r8nd\" (UID: \"f85d16c2-33be-4772-a597-fa4393b5846c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.516227 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.591196 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.597246 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf7952f-0cb5-4873-a116-b878ccdbf7e2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-wtc5f\" (UID: \"cbf7952f-0cb5-4873-a116-b878ccdbf7e2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.674675 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e1b36d1-ed58-439e-ad42-fe17649b18d6" path="/var/lib/kubelet/pods/0e1b36d1-ed58-439e-ad42-fe17649b18d6/volumes" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.688128 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.690549 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f54f5f4f4-xnz7z"] Dec 06 05:54:20 crc kubenswrapper[4957]: W1206 05:54:20.693370 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ee5dadd_4803_4a63_8d41_651780b2d689.slice/crio-f15ddea1442d88fea2ab50a154804ef735de05eb324da2e13687d6d8e5649ea2 WatchSource:0}: Error finding container f15ddea1442d88fea2ab50a154804ef735de05eb324da2e13687d6d8e5649ea2: Status 404 returned error can't find the container with id f15ddea1442d88fea2ab50a154804ef735de05eb324da2e13687d6d8e5649ea2 Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.812152 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f54f5f4f4-xnz7z" event={"ID":"5ee5dadd-4803-4a63-8d41-651780b2d689","Type":"ContainerStarted","Data":"f15ddea1442d88fea2ab50a154804ef735de05eb324da2e13687d6d8e5649ea2"} Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.812683 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-xh9pd" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.815190 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-x8lpv" event={"ID":"211db7e0-522a-4635-a279-1d58502c1d1f","Type":"ContainerStarted","Data":"0c7d61171257390a9c2031a482268a4063b072f93f7c9c612497bccd2a620630"} Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.817784 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" event={"ID":"a60b6144-6a1f-4a86-ae5e-84d5771a8c9c","Type":"ContainerStarted","Data":"9bd3030b9a70f0c24bd0218df61d4e475d7d17234fa942faf46624566ae79f11"} Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.820635 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" Dec 06 05:54:20 crc kubenswrapper[4957]: I1206 05:54:20.923413 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd"] Dec 06 05:54:20 crc kubenswrapper[4957]: W1206 05:54:20.933560 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85d16c2_33be_4772_a597_fa4393b5846c.slice/crio-4b9ded1535846014347e509e09c0ec5db799061310752c396819143ef0710a64 WatchSource:0}: Error finding container 4b9ded1535846014347e509e09c0ec5db799061310752c396819143ef0710a64: Status 404 returned error can't find the container with id 4b9ded1535846014347e509e09c0ec5db799061310752c396819143ef0710a64 Dec 06 05:54:21 crc kubenswrapper[4957]: I1206 05:54:21.026569 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f"] Dec 06 05:54:21 crc kubenswrapper[4957]: W1206 05:54:21.034460 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbf7952f_0cb5_4873_a116_b878ccdbf7e2.slice/crio-17f8893748628a095779d4d0aea51f68e6c991baf78fc9ba2d42867c7aab2fe2 WatchSource:0}: Error finding container 17f8893748628a095779d4d0aea51f68e6c991baf78fc9ba2d42867c7aab2fe2: Status 404 returned error can't find the container with id 17f8893748628a095779d4d0aea51f68e6c991baf78fc9ba2d42867c7aab2fe2 Dec 06 05:54:21 crc kubenswrapper[4957]: I1206 05:54:21.823638 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f54f5f4f4-xnz7z" event={"ID":"5ee5dadd-4803-4a63-8d41-651780b2d689","Type":"ContainerStarted","Data":"c286419a4de5503f2e3dcc969a10627f5a305fd682ea68d6c1951592548bd863"} Dec 06 05:54:21 crc kubenswrapper[4957]: I1206 05:54:21.825520 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zznbp" event={"ID":"2c524bd5-ac23-4921-a527-a6e644e3aa09","Type":"ContainerStarted","Data":"8a76de652f083d7a993593da3f4f51fa329ea5418db22743160e66d18e1cf981"} Dec 06 05:54:21 crc kubenswrapper[4957]: I1206 05:54:21.826849 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" event={"ID":"cbf7952f-0cb5-4873-a116-b878ccdbf7e2","Type":"ContainerStarted","Data":"17f8893748628a095779d4d0aea51f68e6c991baf78fc9ba2d42867c7aab2fe2"} Dec 06 05:54:21 crc kubenswrapper[4957]: I1206 05:54:21.828442 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" event={"ID":"f85d16c2-33be-4772-a597-fa4393b5846c","Type":"ContainerStarted","Data":"4b9ded1535846014347e509e09c0ec5db799061310752c396819143ef0710a64"} Dec 06 05:54:21 crc kubenswrapper[4957]: I1206 05:54:21.844105 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6f54f5f4f4-xnz7z" podStartSLOduration=1.844086795 podStartE2EDuration="1.844086795s" podCreationTimestamp="2025-12-06 05:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:54:21.839726955 +0000 UTC m=+901.489994597" watchObservedRunningTime="2025-12-06 05:54:21.844086795 +0000 UTC m=+901.494354427" Dec 06 05:54:21 crc kubenswrapper[4957]: I1206 05:54:21.866726 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zznbp" podStartSLOduration=2.710843378 podStartE2EDuration="4.866704848s" podCreationTimestamp="2025-12-06 05:54:17 +0000 UTC" firstStartedPulling="2025-12-06 05:54:18.755751502 +0000 UTC m=+898.406019154" lastFinishedPulling="2025-12-06 05:54:20.911612992 +0000 UTC m=+900.561880624" observedRunningTime="2025-12-06 05:54:21.862584594 +0000 UTC m=+901.512852246" watchObservedRunningTime="2025-12-06 05:54:21.866704848 +0000 UTC m=+901.516972470" Dec 06 05:54:27 crc kubenswrapper[4957]: I1206 05:54:27.474867 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:27 crc kubenswrapper[4957]: I1206 05:54:27.477656 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:27 crc kubenswrapper[4957]: I1206 05:54:27.524935 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:27 crc kubenswrapper[4957]: I1206 05:54:27.915181 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:27 crc kubenswrapper[4957]: I1206 05:54:27.953948 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zznbp"] Dec 06 05:54:28 crc kubenswrapper[4957]: I1206 05:54:28.297822 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:54:28 crc kubenswrapper[4957]: I1206 05:54:28.297942 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:54:28 crc kubenswrapper[4957]: I1206 05:54:28.344290 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:54:28 crc kubenswrapper[4957]: I1206 05:54:28.931575 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vsrb4" Dec 06 05:54:29 crc kubenswrapper[4957]: I1206 05:54:29.891941 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vsrb4"] Dec 06 05:54:29 crc kubenswrapper[4957]: I1206 05:54:29.894556 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zznbp" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="registry-server" containerID="cri-o://8a76de652f083d7a993593da3f4f51fa329ea5418db22743160e66d18e1cf981" gracePeriod=2 Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.161435 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wmqv2"] Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.161730 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wmqv2" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="registry-server" containerID="cri-o://08045462e07788e92a2c6ba18daf7194e7416e1a373b015b135233db78000288" gracePeriod=2 Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.517075 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.517123 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.521746 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.901242 4957 generic.go:334] "Generic (PLEG): container finished" podID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerID="8a76de652f083d7a993593da3f4f51fa329ea5418db22743160e66d18e1cf981" exitCode=0 Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.901340 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zznbp" event={"ID":"2c524bd5-ac23-4921-a527-a6e644e3aa09","Type":"ContainerDied","Data":"8a76de652f083d7a993593da3f4f51fa329ea5418db22743160e66d18e1cf981"} Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.903792 4957 generic.go:334] "Generic (PLEG): container finished" podID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerID="08045462e07788e92a2c6ba18daf7194e7416e1a373b015b135233db78000288" exitCode=0 Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.903863 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmqv2" event={"ID":"6a8ca263-88ab-45bb-80d4-70add07a7b14","Type":"ContainerDied","Data":"08045462e07788e92a2c6ba18daf7194e7416e1a373b015b135233db78000288"} Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.907036 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6f54f5f4f4-xnz7z" Dec 06 05:54:30 crc kubenswrapper[4957]: I1206 05:54:30.969907 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8mb7l"] Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.123112 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.130118 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.269288 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-utilities\") pod \"6a8ca263-88ab-45bb-80d4-70add07a7b14\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.269339 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pshk9\" (UniqueName: \"kubernetes.io/projected/6a8ca263-88ab-45bb-80d4-70add07a7b14-kube-api-access-pshk9\") pod \"6a8ca263-88ab-45bb-80d4-70add07a7b14\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.269366 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcl6r\" (UniqueName: \"kubernetes.io/projected/2c524bd5-ac23-4921-a527-a6e644e3aa09-kube-api-access-xcl6r\") pod \"2c524bd5-ac23-4921-a527-a6e644e3aa09\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.269484 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-utilities\") pod \"2c524bd5-ac23-4921-a527-a6e644e3aa09\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.269515 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-catalog-content\") pod \"6a8ca263-88ab-45bb-80d4-70add07a7b14\" (UID: \"6a8ca263-88ab-45bb-80d4-70add07a7b14\") " Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.269563 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-catalog-content\") pod \"2c524bd5-ac23-4921-a527-a6e644e3aa09\" (UID: \"2c524bd5-ac23-4921-a527-a6e644e3aa09\") " Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.270431 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-utilities" (OuterVolumeSpecName: "utilities") pod "6a8ca263-88ab-45bb-80d4-70add07a7b14" (UID: "6a8ca263-88ab-45bb-80d4-70add07a7b14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.270544 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-utilities" (OuterVolumeSpecName: "utilities") pod "2c524bd5-ac23-4921-a527-a6e644e3aa09" (UID: "2c524bd5-ac23-4921-a527-a6e644e3aa09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.275515 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c524bd5-ac23-4921-a527-a6e644e3aa09-kube-api-access-xcl6r" (OuterVolumeSpecName: "kube-api-access-xcl6r") pod "2c524bd5-ac23-4921-a527-a6e644e3aa09" (UID: "2c524bd5-ac23-4921-a527-a6e644e3aa09"). InnerVolumeSpecName "kube-api-access-xcl6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.276458 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8ca263-88ab-45bb-80d4-70add07a7b14-kube-api-access-pshk9" (OuterVolumeSpecName: "kube-api-access-pshk9") pod "6a8ca263-88ab-45bb-80d4-70add07a7b14" (UID: "6a8ca263-88ab-45bb-80d4-70add07a7b14"). InnerVolumeSpecName "kube-api-access-pshk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.293994 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c524bd5-ac23-4921-a527-a6e644e3aa09" (UID: "2c524bd5-ac23-4921-a527-a6e644e3aa09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.319751 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a8ca263-88ab-45bb-80d4-70add07a7b14" (UID: "6a8ca263-88ab-45bb-80d4-70add07a7b14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.371277 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.371341 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.371357 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c524bd5-ac23-4921-a527-a6e644e3aa09-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.371369 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8ca263-88ab-45bb-80d4-70add07a7b14-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.371401 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pshk9\" (UniqueName: \"kubernetes.io/projected/6a8ca263-88ab-45bb-80d4-70add07a7b14-kube-api-access-pshk9\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.371415 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcl6r\" (UniqueName: \"kubernetes.io/projected/2c524bd5-ac23-4921-a527-a6e644e3aa09-kube-api-access-xcl6r\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.927799 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zznbp" event={"ID":"2c524bd5-ac23-4921-a527-a6e644e3aa09","Type":"ContainerDied","Data":"1a5fa559f50888b573ae91a97ec67509e6fb8c461fcf45067fb829eda17067ab"} Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.928252 4957 scope.go:117] "RemoveContainer" containerID="8a76de652f083d7a993593da3f4f51fa329ea5418db22743160e66d18e1cf981" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.928026 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zznbp" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.934177 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmqv2" event={"ID":"6a8ca263-88ab-45bb-80d4-70add07a7b14","Type":"ContainerDied","Data":"e48c441b1f04072e291c583ece35897528c3336ec1f8475a6fec304e36f0aa84"} Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.934286 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmqv2" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.951654 4957 scope.go:117] "RemoveContainer" containerID="2b8fbe80034c7602898fd1afffffe82d18ad3adc9f262feec00ec340679fd3f3" Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.962925 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zznbp"] Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.967302 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zznbp"] Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.975888 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wmqv2"] Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.980054 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wmqv2"] Dec 06 05:54:32 crc kubenswrapper[4957]: I1206 05:54:32.991698 4957 scope.go:117] "RemoveContainer" containerID="f90dba1cdd20d8b0584e55c9a5f99f9299893fcdb710762e1819b4761ee16ba9" Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.008977 4957 scope.go:117] "RemoveContainer" containerID="08045462e07788e92a2c6ba18daf7194e7416e1a373b015b135233db78000288" Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.023361 4957 scope.go:117] "RemoveContainer" containerID="04a719f3864b72821b857ef55dc6d96f190ab45597c647381b46c1d052f8dbf7" Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.039101 4957 scope.go:117] "RemoveContainer" containerID="305fd7f55e6d3ec6aa50dfb8f828124b00513db55ab707201233f4f13b224087" Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.943601 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" event={"ID":"cbf7952f-0cb5-4873-a116-b878ccdbf7e2","Type":"ContainerStarted","Data":"206e03a39e69b29b60957897d5ad5d1d71b27565f8de3b08423f5a98a74eb806"} Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.945147 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" event={"ID":"f85d16c2-33be-4772-a597-fa4393b5846c","Type":"ContainerStarted","Data":"b1d2dc20aa8f55c0535997fb550b8b73899b585e1a5ef51c8de38b2b88b14ada"} Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.945538 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.948332 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" event={"ID":"a60b6144-6a1f-4a86-ae5e-84d5771a8c9c","Type":"ContainerStarted","Data":"df056376e44b6e929da3b2f269e541c0293779b7b4de18a726dde2d7aa44fcc3"} Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.952118 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-x8lpv" event={"ID":"211db7e0-522a-4635-a279-1d58502c1d1f","Type":"ContainerStarted","Data":"3ca1b352fc4f936c3dfdabc67db7bf097818ac661e49c287c2322c8b971fc4d4"} Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.952391 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.969872 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-wtc5f" podStartSLOduration=3.598357445 podStartE2EDuration="14.969814814s" podCreationTimestamp="2025-12-06 05:54:19 +0000 UTC" firstStartedPulling="2025-12-06 05:54:21.038966519 +0000 UTC m=+900.689234151" lastFinishedPulling="2025-12-06 05:54:32.410423888 +0000 UTC m=+912.060691520" observedRunningTime="2025-12-06 05:54:33.962482721 +0000 UTC m=+913.612750373" watchObservedRunningTime="2025-12-06 05:54:33.969814814 +0000 UTC m=+913.620082446" Dec 06 05:54:33 crc kubenswrapper[4957]: I1206 05:54:33.990771 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" podStartSLOduration=3.4275058769999998 podStartE2EDuration="14.990747105s" podCreationTimestamp="2025-12-06 05:54:19 +0000 UTC" firstStartedPulling="2025-12-06 05:54:20.9362262 +0000 UTC m=+900.586493832" lastFinishedPulling="2025-12-06 05:54:32.499467428 +0000 UTC m=+912.149735060" observedRunningTime="2025-12-06 05:54:33.984734989 +0000 UTC m=+913.635002631" watchObservedRunningTime="2025-12-06 05:54:33.990747105 +0000 UTC m=+913.641014747" Dec 06 05:54:34 crc kubenswrapper[4957]: I1206 05:54:34.003528 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-x8lpv" podStartSLOduration=2.7678497 podStartE2EDuration="15.003495099s" podCreationTimestamp="2025-12-06 05:54:19 +0000 UTC" firstStartedPulling="2025-12-06 05:54:20.167781484 +0000 UTC m=+899.818049126" lastFinishedPulling="2025-12-06 05:54:32.403426893 +0000 UTC m=+912.053694525" observedRunningTime="2025-12-06 05:54:33.999359574 +0000 UTC m=+913.649627216" watchObservedRunningTime="2025-12-06 05:54:34.003495099 +0000 UTC m=+913.653762731" Dec 06 05:54:34 crc kubenswrapper[4957]: I1206 05:54:34.669415 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" path="/var/lib/kubelet/pods/2c524bd5-ac23-4921-a527-a6e644e3aa09/volumes" Dec 06 05:54:34 crc kubenswrapper[4957]: I1206 05:54:34.670095 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" path="/var/lib/kubelet/pods/6a8ca263-88ab-45bb-80d4-70add07a7b14/volumes" Dec 06 05:54:40 crc kubenswrapper[4957]: I1206 05:54:40.151785 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-x8lpv" Dec 06 05:54:44 crc kubenswrapper[4957]: I1206 05:54:44.013823 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" event={"ID":"a60b6144-6a1f-4a86-ae5e-84d5771a8c9c","Type":"ContainerStarted","Data":"81d458e73af25988e76e3c63d286ea5e0942c3f25f60f7dc72c5698c849bd3d6"} Dec 06 05:54:44 crc kubenswrapper[4957]: I1206 05:54:44.037459 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mlsvn" podStartSLOduration=2.257867389 podStartE2EDuration="25.03744133s" podCreationTimestamp="2025-12-06 05:54:19 +0000 UTC" firstStartedPulling="2025-12-06 05:54:20.393383198 +0000 UTC m=+900.043650830" lastFinishedPulling="2025-12-06 05:54:43.172957139 +0000 UTC m=+922.823224771" observedRunningTime="2025-12-06 05:54:44.03671197 +0000 UTC m=+923.686979652" watchObservedRunningTime="2025-12-06 05:54:44.03744133 +0000 UTC m=+923.687708972" Dec 06 05:54:50 crc kubenswrapper[4957]: I1206 05:54:50.692914 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-6r8nd" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.024285 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8mb7l" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerName="console" containerID="cri-o://810ab1c8eead3866218d55cf51d0eb564320b73776026fdd3bc9cf753fe1b3e4" gracePeriod=15 Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414307 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xfkmq"] Dec 06 05:54:56 crc kubenswrapper[4957]: E1206 05:54:56.414579 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="extract-utilities" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414593 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="extract-utilities" Dec 06 05:54:56 crc kubenswrapper[4957]: E1206 05:54:56.414602 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="extract-content" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414608 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="extract-content" Dec 06 05:54:56 crc kubenswrapper[4957]: E1206 05:54:56.414619 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="extract-utilities" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414626 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="extract-utilities" Dec 06 05:54:56 crc kubenswrapper[4957]: E1206 05:54:56.414641 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="extract-content" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414648 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="extract-content" Dec 06 05:54:56 crc kubenswrapper[4957]: E1206 05:54:56.414659 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="registry-server" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414666 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="registry-server" Dec 06 05:54:56 crc kubenswrapper[4957]: E1206 05:54:56.414674 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="registry-server" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414681 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="registry-server" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414852 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c524bd5-ac23-4921-a527-a6e644e3aa09" containerName="registry-server" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.414870 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8ca263-88ab-45bb-80d4-70add07a7b14" containerName="registry-server" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.417602 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.438357 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfkmq"] Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.554581 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ztsg\" (UniqueName: \"kubernetes.io/projected/9dad819b-e7a2-4565-9018-5dbadec08bfa-kube-api-access-4ztsg\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.554755 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-catalog-content\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.554872 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-utilities\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.655970 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-catalog-content\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.656193 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-utilities\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.656242 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ztsg\" (UniqueName: \"kubernetes.io/projected/9dad819b-e7a2-4565-9018-5dbadec08bfa-kube-api-access-4ztsg\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.657011 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-catalog-content\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.657221 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-utilities\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.679884 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ztsg\" (UniqueName: \"kubernetes.io/projected/9dad819b-e7a2-4565-9018-5dbadec08bfa-kube-api-access-4ztsg\") pod \"certified-operators-xfkmq\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.707628 4957 patch_prober.go:28] interesting pod/console-f9d7485db-8mb7l container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/health\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.707685 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-8mb7l" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerName="console" probeResult="failure" output="Get \"https://10.217.0.37:8443/health\": dial tcp 10.217.0.37:8443: connect: connection refused" Dec 06 05:54:56 crc kubenswrapper[4957]: I1206 05:54:56.737626 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.088918 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8mb7l_0483ff2c-b2e8-44f2-bec5-85eb4d06d279/console/0.log" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.089138 4957 generic.go:334] "Generic (PLEG): container finished" podID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerID="810ab1c8eead3866218d55cf51d0eb564320b73776026fdd3bc9cf753fe1b3e4" exitCode=2 Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.089168 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8mb7l" event={"ID":"0483ff2c-b2e8-44f2-bec5-85eb4d06d279","Type":"ContainerDied","Data":"810ab1c8eead3866218d55cf51d0eb564320b73776026fdd3bc9cf753fe1b3e4"} Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.238032 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfkmq"] Dec 06 05:54:57 crc kubenswrapper[4957]: W1206 05:54:57.241718 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dad819b_e7a2_4565_9018_5dbadec08bfa.slice/crio-c4ed4d218525d9728854a8f067b4e67a19333d454fb09f29a4bba79cb5e6b8c0 WatchSource:0}: Error finding container c4ed4d218525d9728854a8f067b4e67a19333d454fb09f29a4bba79cb5e6b8c0: Status 404 returned error can't find the container with id c4ed4d218525d9728854a8f067b4e67a19333d454fb09f29a4bba79cb5e6b8c0 Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.505235 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8mb7l_0483ff2c-b2e8-44f2-bec5-85eb4d06d279/console/0.log" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.505521 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.574707 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-config\") pod \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.574772 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-serving-cert\") pod \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.574876 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-oauth-config\") pod \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.574921 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-service-ca\") pod \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.574940 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swrrj\" (UniqueName: \"kubernetes.io/projected/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-kube-api-access-swrrj\") pod \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.574961 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-trusted-ca-bundle\") pod \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.574983 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-oauth-serving-cert\") pod \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\" (UID: \"0483ff2c-b2e8-44f2-bec5-85eb4d06d279\") " Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.576585 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-service-ca" (OuterVolumeSpecName: "service-ca") pod "0483ff2c-b2e8-44f2-bec5-85eb4d06d279" (UID: "0483ff2c-b2e8-44f2-bec5-85eb4d06d279"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.576970 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-config" (OuterVolumeSpecName: "console-config") pod "0483ff2c-b2e8-44f2-bec5-85eb4d06d279" (UID: "0483ff2c-b2e8-44f2-bec5-85eb4d06d279"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.577210 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0483ff2c-b2e8-44f2-bec5-85eb4d06d279" (UID: "0483ff2c-b2e8-44f2-bec5-85eb4d06d279"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.577293 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0483ff2c-b2e8-44f2-bec5-85eb4d06d279" (UID: "0483ff2c-b2e8-44f2-bec5-85eb4d06d279"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.580901 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0483ff2c-b2e8-44f2-bec5-85eb4d06d279" (UID: "0483ff2c-b2e8-44f2-bec5-85eb4d06d279"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.580976 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-kube-api-access-swrrj" (OuterVolumeSpecName: "kube-api-access-swrrj") pod "0483ff2c-b2e8-44f2-bec5-85eb4d06d279" (UID: "0483ff2c-b2e8-44f2-bec5-85eb4d06d279"). InnerVolumeSpecName "kube-api-access-swrrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.581151 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0483ff2c-b2e8-44f2-bec5-85eb4d06d279" (UID: "0483ff2c-b2e8-44f2-bec5-85eb4d06d279"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.677014 4957 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.678274 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swrrj\" (UniqueName: \"kubernetes.io/projected/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-kube-api-access-swrrj\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.678290 4957 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.678303 4957 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.678314 4957 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.678325 4957 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:57 crc kubenswrapper[4957]: I1206 05:54:57.678339 4957 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0483ff2c-b2e8-44f2-bec5-85eb4d06d279-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.100923 4957 generic.go:334] "Generic (PLEG): container finished" podID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerID="8dbde4e987a442168d6bfd4bf6a716cef0e338dce36287b1e7fbd462dbdcfd48" exitCode=0 Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.101222 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfkmq" event={"ID":"9dad819b-e7a2-4565-9018-5dbadec08bfa","Type":"ContainerDied","Data":"8dbde4e987a442168d6bfd4bf6a716cef0e338dce36287b1e7fbd462dbdcfd48"} Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.102235 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfkmq" event={"ID":"9dad819b-e7a2-4565-9018-5dbadec08bfa","Type":"ContainerStarted","Data":"c4ed4d218525d9728854a8f067b4e67a19333d454fb09f29a4bba79cb5e6b8c0"} Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.105675 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8mb7l_0483ff2c-b2e8-44f2-bec5-85eb4d06d279/console/0.log" Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.105791 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8mb7l" event={"ID":"0483ff2c-b2e8-44f2-bec5-85eb4d06d279","Type":"ContainerDied","Data":"a6fc479e6daf88890e0bc7d5738329117bf7a04d752a649ab872dd5f47e2ac9c"} Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.105902 4957 scope.go:117] "RemoveContainer" containerID="810ab1c8eead3866218d55cf51d0eb564320b73776026fdd3bc9cf753fe1b3e4" Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.106077 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8mb7l" Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.150957 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8mb7l"] Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.152363 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8mb7l"] Dec 06 05:54:58 crc kubenswrapper[4957]: I1206 05:54:58.671247 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" path="/var/lib/kubelet/pods/0483ff2c-b2e8-44f2-bec5-85eb4d06d279/volumes" Dec 06 05:54:59 crc kubenswrapper[4957]: I1206 05:54:59.120997 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfkmq" event={"ID":"9dad819b-e7a2-4565-9018-5dbadec08bfa","Type":"ContainerStarted","Data":"bfa79c954437f7ad28dfec6a9af087d8235b07e0d355604bd2fce9c387d3790f"} Dec 06 05:55:00 crc kubenswrapper[4957]: I1206 05:55:00.126754 4957 generic.go:334] "Generic (PLEG): container finished" podID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerID="bfa79c954437f7ad28dfec6a9af087d8235b07e0d355604bd2fce9c387d3790f" exitCode=0 Dec 06 05:55:00 crc kubenswrapper[4957]: I1206 05:55:00.127016 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfkmq" event={"ID":"9dad819b-e7a2-4565-9018-5dbadec08bfa","Type":"ContainerDied","Data":"bfa79c954437f7ad28dfec6a9af087d8235b07e0d355604bd2fce9c387d3790f"} Dec 06 05:55:01 crc kubenswrapper[4957]: I1206 05:55:01.136572 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfkmq" event={"ID":"9dad819b-e7a2-4565-9018-5dbadec08bfa","Type":"ContainerStarted","Data":"8aa6be3ec80f81d0ad7fcac6808838c84a768ca14fb5ff2581e3d96e07ed43f9"} Dec 06 05:55:01 crc kubenswrapper[4957]: I1206 05:55:01.153748 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xfkmq" podStartSLOduration=2.764257407 podStartE2EDuration="5.153663388s" podCreationTimestamp="2025-12-06 05:54:56 +0000 UTC" firstStartedPulling="2025-12-06 05:54:58.1038833 +0000 UTC m=+937.754150922" lastFinishedPulling="2025-12-06 05:55:00.493289261 +0000 UTC m=+940.143556903" observedRunningTime="2025-12-06 05:55:01.150934202 +0000 UTC m=+940.801201844" watchObservedRunningTime="2025-12-06 05:55:01.153663388 +0000 UTC m=+940.803931020" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.046618 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b"] Dec 06 05:55:04 crc kubenswrapper[4957]: E1206 05:55:04.047394 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerName="console" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.047406 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerName="console" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.047515 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="0483ff2c-b2e8-44f2-bec5-85eb4d06d279" containerName="console" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.048241 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.050293 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.056690 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b"] Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.082528 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.082573 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.082633 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45scb\" (UniqueName: \"kubernetes.io/projected/555b7328-82f2-4233-ba2a-de0145c1560b-kube-api-access-45scb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.185051 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45scb\" (UniqueName: \"kubernetes.io/projected/555b7328-82f2-4233-ba2a-de0145c1560b-kube-api-access-45scb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.185633 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.185676 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.186900 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.187017 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.204011 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45scb\" (UniqueName: \"kubernetes.io/projected/555b7328-82f2-4233-ba2a-de0145c1560b-kube-api-access-45scb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.392552 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:04 crc kubenswrapper[4957]: I1206 05:55:04.656554 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b"] Dec 06 05:55:05 crc kubenswrapper[4957]: I1206 05:55:05.170542 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" event={"ID":"555b7328-82f2-4233-ba2a-de0145c1560b","Type":"ContainerStarted","Data":"db99e6c84bc9d799cb452b67ae706bd0bbd05b67b3fffb043588fb307a0ec835"} Dec 06 05:55:06 crc kubenswrapper[4957]: I1206 05:55:06.176939 4957 generic.go:334] "Generic (PLEG): container finished" podID="555b7328-82f2-4233-ba2a-de0145c1560b" containerID="22536a192e3e0f5032e2256b4e09e080e4b669de0cbb33063520b903ab050eda" exitCode=0 Dec 06 05:55:06 crc kubenswrapper[4957]: I1206 05:55:06.177031 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" event={"ID":"555b7328-82f2-4233-ba2a-de0145c1560b","Type":"ContainerDied","Data":"22536a192e3e0f5032e2256b4e09e080e4b669de0cbb33063520b903ab050eda"} Dec 06 05:55:06 crc kubenswrapper[4957]: I1206 05:55:06.738886 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:55:06 crc kubenswrapper[4957]: I1206 05:55:06.738964 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:55:06 crc kubenswrapper[4957]: I1206 05:55:06.775849 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:55:07 crc kubenswrapper[4957]: I1206 05:55:07.232032 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:55:08 crc kubenswrapper[4957]: I1206 05:55:08.997006 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfkmq"] Dec 06 05:55:09 crc kubenswrapper[4957]: I1206 05:55:09.194323 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xfkmq" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="registry-server" containerID="cri-o://8aa6be3ec80f81d0ad7fcac6808838c84a768ca14fb5ff2581e3d96e07ed43f9" gracePeriod=2 Dec 06 05:55:10 crc kubenswrapper[4957]: I1206 05:55:10.180769 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:55:10 crc kubenswrapper[4957]: I1206 05:55:10.181141 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:55:10 crc kubenswrapper[4957]: I1206 05:55:10.200353 4957 generic.go:334] "Generic (PLEG): container finished" podID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerID="8aa6be3ec80f81d0ad7fcac6808838c84a768ca14fb5ff2581e3d96e07ed43f9" exitCode=0 Dec 06 05:55:10 crc kubenswrapper[4957]: I1206 05:55:10.200454 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfkmq" event={"ID":"9dad819b-e7a2-4565-9018-5dbadec08bfa","Type":"ContainerDied","Data":"8aa6be3ec80f81d0ad7fcac6808838c84a768ca14fb5ff2581e3d96e07ed43f9"} Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.755103 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.829726 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ztsg\" (UniqueName: \"kubernetes.io/projected/9dad819b-e7a2-4565-9018-5dbadec08bfa-kube-api-access-4ztsg\") pod \"9dad819b-e7a2-4565-9018-5dbadec08bfa\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.829803 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-utilities\") pod \"9dad819b-e7a2-4565-9018-5dbadec08bfa\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.829878 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-catalog-content\") pod \"9dad819b-e7a2-4565-9018-5dbadec08bfa\" (UID: \"9dad819b-e7a2-4565-9018-5dbadec08bfa\") " Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.830826 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-utilities" (OuterVolumeSpecName: "utilities") pod "9dad819b-e7a2-4565-9018-5dbadec08bfa" (UID: "9dad819b-e7a2-4565-9018-5dbadec08bfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.843089 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dad819b-e7a2-4565-9018-5dbadec08bfa-kube-api-access-4ztsg" (OuterVolumeSpecName: "kube-api-access-4ztsg") pod "9dad819b-e7a2-4565-9018-5dbadec08bfa" (UID: "9dad819b-e7a2-4565-9018-5dbadec08bfa"). InnerVolumeSpecName "kube-api-access-4ztsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.877913 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9dad819b-e7a2-4565-9018-5dbadec08bfa" (UID: "9dad819b-e7a2-4565-9018-5dbadec08bfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.931754 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ztsg\" (UniqueName: \"kubernetes.io/projected/9dad819b-e7a2-4565-9018-5dbadec08bfa-kube-api-access-4ztsg\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.931788 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:14 crc kubenswrapper[4957]: I1206 05:55:14.931797 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dad819b-e7a2-4565-9018-5dbadec08bfa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.235878 4957 generic.go:334] "Generic (PLEG): container finished" podID="555b7328-82f2-4233-ba2a-de0145c1560b" containerID="3db1e2037458e9b3f0233154018657f52f0831a7a3dcf137ff5057c97f845ed0" exitCode=0 Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.236011 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" event={"ID":"555b7328-82f2-4233-ba2a-de0145c1560b","Type":"ContainerDied","Data":"3db1e2037458e9b3f0233154018657f52f0831a7a3dcf137ff5057c97f845ed0"} Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.241277 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfkmq" event={"ID":"9dad819b-e7a2-4565-9018-5dbadec08bfa","Type":"ContainerDied","Data":"c4ed4d218525d9728854a8f067b4e67a19333d454fb09f29a4bba79cb5e6b8c0"} Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.241346 4957 scope.go:117] "RemoveContainer" containerID="8aa6be3ec80f81d0ad7fcac6808838c84a768ca14fb5ff2581e3d96e07ed43f9" Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.241543 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfkmq" Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.305898 4957 scope.go:117] "RemoveContainer" containerID="bfa79c954437f7ad28dfec6a9af087d8235b07e0d355604bd2fce9c387d3790f" Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.330490 4957 scope.go:117] "RemoveContainer" containerID="8dbde4e987a442168d6bfd4bf6a716cef0e338dce36287b1e7fbd462dbdcfd48" Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.359375 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfkmq"] Dec 06 05:55:15 crc kubenswrapper[4957]: I1206 05:55:15.362793 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xfkmq"] Dec 06 05:55:16 crc kubenswrapper[4957]: I1206 05:55:16.253403 4957 generic.go:334] "Generic (PLEG): container finished" podID="555b7328-82f2-4233-ba2a-de0145c1560b" containerID="8901d0e3e3be17b1379a94460bd7a1e9ee8ce127e4bb57e8e960fc6edf9a38e5" exitCode=0 Dec 06 05:55:16 crc kubenswrapper[4957]: I1206 05:55:16.253479 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" event={"ID":"555b7328-82f2-4233-ba2a-de0145c1560b","Type":"ContainerDied","Data":"8901d0e3e3be17b1379a94460bd7a1e9ee8ce127e4bb57e8e960fc6edf9a38e5"} Dec 06 05:55:16 crc kubenswrapper[4957]: I1206 05:55:16.669174 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" path="/var/lib/kubelet/pods/9dad819b-e7a2-4565-9018-5dbadec08bfa/volumes" Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.467357 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.667286 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-util\") pod \"555b7328-82f2-4233-ba2a-de0145c1560b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.667341 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45scb\" (UniqueName: \"kubernetes.io/projected/555b7328-82f2-4233-ba2a-de0145c1560b-kube-api-access-45scb\") pod \"555b7328-82f2-4233-ba2a-de0145c1560b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.667423 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-bundle\") pod \"555b7328-82f2-4233-ba2a-de0145c1560b\" (UID: \"555b7328-82f2-4233-ba2a-de0145c1560b\") " Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.668412 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-bundle" (OuterVolumeSpecName: "bundle") pod "555b7328-82f2-4233-ba2a-de0145c1560b" (UID: "555b7328-82f2-4233-ba2a-de0145c1560b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.672322 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555b7328-82f2-4233-ba2a-de0145c1560b-kube-api-access-45scb" (OuterVolumeSpecName: "kube-api-access-45scb") pod "555b7328-82f2-4233-ba2a-de0145c1560b" (UID: "555b7328-82f2-4233-ba2a-de0145c1560b"). InnerVolumeSpecName "kube-api-access-45scb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.681091 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-util" (OuterVolumeSpecName: "util") pod "555b7328-82f2-4233-ba2a-de0145c1560b" (UID: "555b7328-82f2-4233-ba2a-de0145c1560b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.768775 4957 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-util\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.768820 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45scb\" (UniqueName: \"kubernetes.io/projected/555b7328-82f2-4233-ba2a-de0145c1560b-kube-api-access-45scb\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:17 crc kubenswrapper[4957]: I1206 05:55:17.768893 4957 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/555b7328-82f2-4233-ba2a-de0145c1560b-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:55:18 crc kubenswrapper[4957]: I1206 05:55:18.264627 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" event={"ID":"555b7328-82f2-4233-ba2a-de0145c1560b","Type":"ContainerDied","Data":"db99e6c84bc9d799cb452b67ae706bd0bbd05b67b3fffb043588fb307a0ec835"} Dec 06 05:55:18 crc kubenswrapper[4957]: I1206 05:55:18.264668 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db99e6c84bc9d799cb452b67ae706bd0bbd05b67b3fffb043588fb307a0ec835" Dec 06 05:55:18 crc kubenswrapper[4957]: I1206 05:55:18.264792 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.299553 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm"] Dec 06 05:55:26 crc kubenswrapper[4957]: E1206 05:55:26.300424 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555b7328-82f2-4233-ba2a-de0145c1560b" containerName="extract" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.300439 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="555b7328-82f2-4233-ba2a-de0145c1560b" containerName="extract" Dec 06 05:55:26 crc kubenswrapper[4957]: E1206 05:55:26.300452 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="extract-content" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.300458 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="extract-content" Dec 06 05:55:26 crc kubenswrapper[4957]: E1206 05:55:26.300477 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="registry-server" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.300484 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="registry-server" Dec 06 05:55:26 crc kubenswrapper[4957]: E1206 05:55:26.300491 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="extract-utilities" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.300496 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="extract-utilities" Dec 06 05:55:26 crc kubenswrapper[4957]: E1206 05:55:26.300505 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555b7328-82f2-4233-ba2a-de0145c1560b" containerName="util" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.300511 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="555b7328-82f2-4233-ba2a-de0145c1560b" containerName="util" Dec 06 05:55:26 crc kubenswrapper[4957]: E1206 05:55:26.300521 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555b7328-82f2-4233-ba2a-de0145c1560b" containerName="pull" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.300526 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="555b7328-82f2-4233-ba2a-de0145c1560b" containerName="pull" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.302164 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dad819b-e7a2-4565-9018-5dbadec08bfa" containerName="registry-server" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.302204 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="555b7328-82f2-4233-ba2a-de0145c1560b" containerName="extract" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.302796 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.305049 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.305135 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.305461 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.305778 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.307238 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-2xv74" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.324595 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm"] Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.410271 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-apiservice-cert\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.410348 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28pw5\" (UniqueName: \"kubernetes.io/projected/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-kube-api-access-28pw5\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.410481 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-webhook-cert\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.511274 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-apiservice-cert\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.511340 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28pw5\" (UniqueName: \"kubernetes.io/projected/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-kube-api-access-28pw5\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.511374 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-webhook-cert\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.517253 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-apiservice-cert\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.521569 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-webhook-cert\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.561112 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28pw5\" (UniqueName: \"kubernetes.io/projected/c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f-kube-api-access-28pw5\") pod \"metallb-operator-controller-manager-9b6b64df4-mn6tm\" (UID: \"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f\") " pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.618163 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl"] Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.619087 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.619893 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.621970 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.621995 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.623943 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-99hcs" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.698043 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl"] Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.714353 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c3dd44fb-8340-44f9-af68-f7e628a59b6c-webhook-cert\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.714435 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c3dd44fb-8340-44f9-af68-f7e628a59b6c-apiservice-cert\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.714587 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfvv5\" (UniqueName: \"kubernetes.io/projected/c3dd44fb-8340-44f9-af68-f7e628a59b6c-kube-api-access-wfvv5\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.815387 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfvv5\" (UniqueName: \"kubernetes.io/projected/c3dd44fb-8340-44f9-af68-f7e628a59b6c-kube-api-access-wfvv5\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.815870 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c3dd44fb-8340-44f9-af68-f7e628a59b6c-webhook-cert\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.816748 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c3dd44fb-8340-44f9-af68-f7e628a59b6c-apiservice-cert\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.819909 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c3dd44fb-8340-44f9-af68-f7e628a59b6c-webhook-cert\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.820381 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c3dd44fb-8340-44f9-af68-f7e628a59b6c-apiservice-cert\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.838633 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfvv5\" (UniqueName: \"kubernetes.io/projected/c3dd44fb-8340-44f9-af68-f7e628a59b6c-kube-api-access-wfvv5\") pod \"metallb-operator-webhook-server-5568f7b8bc-8wsfl\" (UID: \"c3dd44fb-8340-44f9-af68-f7e628a59b6c\") " pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.924477 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm"] Dec 06 05:55:26 crc kubenswrapper[4957]: I1206 05:55:26.988374 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:27 crc kubenswrapper[4957]: I1206 05:55:27.307706 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" event={"ID":"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f","Type":"ContainerStarted","Data":"1c9c7cc34654d29243934c69c6b562d3dd355907da12173c312836583ff856bf"} Dec 06 05:55:27 crc kubenswrapper[4957]: I1206 05:55:27.533726 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl"] Dec 06 05:55:27 crc kubenswrapper[4957]: W1206 05:55:27.538967 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3dd44fb_8340_44f9_af68_f7e628a59b6c.slice/crio-3c656873d2f7f9e7f30b1704458d24158656d78242c7a15a315565ee7dca5493 WatchSource:0}: Error finding container 3c656873d2f7f9e7f30b1704458d24158656d78242c7a15a315565ee7dca5493: Status 404 returned error can't find the container with id 3c656873d2f7f9e7f30b1704458d24158656d78242c7a15a315565ee7dca5493 Dec 06 05:55:28 crc kubenswrapper[4957]: I1206 05:55:28.314504 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" event={"ID":"c3dd44fb-8340-44f9-af68-f7e628a59b6c","Type":"ContainerStarted","Data":"3c656873d2f7f9e7f30b1704458d24158656d78242c7a15a315565ee7dca5493"} Dec 06 05:55:39 crc kubenswrapper[4957]: I1206 05:55:39.369640 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" event={"ID":"c3dd44fb-8340-44f9-af68-f7e628a59b6c","Type":"ContainerStarted","Data":"9d1ac4d7b247e4c07ebff66e8b81fa9902c0b1e808b35cf742d62bad399b4d03"} Dec 06 05:55:39 crc kubenswrapper[4957]: I1206 05:55:39.371367 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" event={"ID":"c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f","Type":"ContainerStarted","Data":"f17f535b2da6b7efb8243c1334ca7cb7dd9f73a2c470226115d6dc252608a9aa"} Dec 06 05:55:39 crc kubenswrapper[4957]: I1206 05:55:39.372272 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:55:40 crc kubenswrapper[4957]: I1206 05:55:40.180481 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:55:40 crc kubenswrapper[4957]: I1206 05:55:40.180791 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:55:40 crc kubenswrapper[4957]: I1206 05:55:40.404196 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" podStartSLOduration=2.828255981 podStartE2EDuration="14.404178919s" podCreationTimestamp="2025-12-06 05:55:26 +0000 UTC" firstStartedPulling="2025-12-06 05:55:26.942026745 +0000 UTC m=+966.592294377" lastFinishedPulling="2025-12-06 05:55:38.517949673 +0000 UTC m=+978.168217315" observedRunningTime="2025-12-06 05:55:39.402190699 +0000 UTC m=+979.052458351" watchObservedRunningTime="2025-12-06 05:55:40.404178919 +0000 UTC m=+980.054446551" Dec 06 05:55:46 crc kubenswrapper[4957]: I1206 05:55:46.988909 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:56 crc kubenswrapper[4957]: I1206 05:55:56.998231 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" Dec 06 05:55:57 crc kubenswrapper[4957]: I1206 05:55:57.020147 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5568f7b8bc-8wsfl" podStartSLOduration=20.024890388 podStartE2EDuration="31.02012837s" podCreationTimestamp="2025-12-06 05:55:26 +0000 UTC" firstStartedPulling="2025-12-06 05:55:27.542492089 +0000 UTC m=+967.192759731" lastFinishedPulling="2025-12-06 05:55:38.537730081 +0000 UTC m=+978.187997713" observedRunningTime="2025-12-06 05:55:40.405914547 +0000 UTC m=+980.056182179" watchObservedRunningTime="2025-12-06 05:55:57.02012837 +0000 UTC m=+996.670396002" Dec 06 05:56:10 crc kubenswrapper[4957]: I1206 05:56:10.180646 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:56:10 crc kubenswrapper[4957]: I1206 05:56:10.181208 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:56:10 crc kubenswrapper[4957]: I1206 05:56:10.181255 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:56:10 crc kubenswrapper[4957]: I1206 05:56:10.181942 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c315dd48c0a384ca90d94b9ce0d09121f3d64da936f2696223ce8145d1b5a516"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:56:10 crc kubenswrapper[4957]: I1206 05:56:10.181996 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://c315dd48c0a384ca90d94b9ce0d09121f3d64da936f2696223ce8145d1b5a516" gracePeriod=600 Dec 06 05:56:11 crc kubenswrapper[4957]: I1206 05:56:11.714762 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="c315dd48c0a384ca90d94b9ce0d09121f3d64da936f2696223ce8145d1b5a516" exitCode=0 Dec 06 05:56:11 crc kubenswrapper[4957]: I1206 05:56:11.714882 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"c315dd48c0a384ca90d94b9ce0d09121f3d64da936f2696223ce8145d1b5a516"} Dec 06 05:56:11 crc kubenswrapper[4957]: I1206 05:56:11.715363 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"f3f75ddbe4ecf31b2a5519fb70a8ed4815614284758f53e65c5fc0533e95c0a3"} Dec 06 05:56:11 crc kubenswrapper[4957]: I1206 05:56:11.715411 4957 scope.go:117] "RemoveContainer" containerID="0b0c4f58bdcf5290181ec1a457cbeb02772611a1bd68d33a39e309f1aaee0927" Dec 06 05:56:16 crc kubenswrapper[4957]: I1206 05:56:16.622754 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-9b6b64df4-mn6tm" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.378748 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p"] Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.380062 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.384013 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.389059 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-wj8pj" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.393269 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-mv448"] Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.395646 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.398195 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.399030 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.400075 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p"] Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.458565 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-metrics\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.458617 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr29w\" (UniqueName: \"kubernetes.io/projected/33816fc1-892d-45d9-b755-803b1bf561bf-kube-api-access-jr29w\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.458688 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5338fe04-633c-405b-8079-1001859b804f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sbt5p\" (UID: \"5338fe04-633c-405b-8079-1001859b804f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.458753 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88kvk\" (UniqueName: \"kubernetes.io/projected/5338fe04-633c-405b-8079-1001859b804f-kube-api-access-88kvk\") pod \"frr-k8s-webhook-server-7fcb986d4-sbt5p\" (UID: \"5338fe04-633c-405b-8079-1001859b804f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.458805 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-reloader\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.458910 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33816fc1-892d-45d9-b755-803b1bf561bf-metrics-certs\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.458974 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-frr-sockets\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.459025 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-frr-conf\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.459070 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33816fc1-892d-45d9-b755-803b1bf561bf-frr-startup\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.492476 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-2jbdq"] Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.493482 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.500629 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.500669 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.501558 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.508362 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-k7t5v" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.533690 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-qzfkk"] Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.534783 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.543095 4957 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559755 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-frr-sockets\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559797 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-frr-conf\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559843 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33816fc1-892d-45d9-b755-803b1bf561bf-frr-startup\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559871 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-metrics\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559887 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr29w\" (UniqueName: \"kubernetes.io/projected/33816fc1-892d-45d9-b755-803b1bf561bf-kube-api-access-jr29w\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559915 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-metrics-certs\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559936 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5338fe04-633c-405b-8079-1001859b804f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sbt5p\" (UID: \"5338fe04-633c-405b-8079-1001859b804f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559957 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88kvk\" (UniqueName: \"kubernetes.io/projected/5338fe04-633c-405b-8079-1001859b804f-kube-api-access-88kvk\") pod \"frr-k8s-webhook-server-7fcb986d4-sbt5p\" (UID: \"5338fe04-633c-405b-8079-1001859b804f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.559986 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-reloader\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560009 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzvnp\" (UniqueName: \"kubernetes.io/projected/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-kube-api-access-fzvnp\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560033 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl5hn\" (UniqueName: \"kubernetes.io/projected/fa194652-baa4-424b-bcd0-736ac7295ec4-kube-api-access-zl5hn\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560056 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-metrics-certs\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560082 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560103 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33816fc1-892d-45d9-b755-803b1bf561bf-metrics-certs\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560122 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-cert\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560139 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fa194652-baa4-424b-bcd0-736ac7295ec4-metallb-excludel2\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560304 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-frr-sockets\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560461 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-frr-conf\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560569 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-metrics\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.563596 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/33816fc1-892d-45d9-b755-803b1bf561bf-frr-startup\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.560583 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/33816fc1-892d-45d9-b755-803b1bf561bf-reloader\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.572912 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5338fe04-633c-405b-8079-1001859b804f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sbt5p\" (UID: \"5338fe04-633c-405b-8079-1001859b804f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.592488 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/33816fc1-892d-45d9-b755-803b1bf561bf-metrics-certs\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.604185 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-qzfkk"] Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.630577 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr29w\" (UniqueName: \"kubernetes.io/projected/33816fc1-892d-45d9-b755-803b1bf561bf-kube-api-access-jr29w\") pod \"frr-k8s-mv448\" (UID: \"33816fc1-892d-45d9-b755-803b1bf561bf\") " pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.635593 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88kvk\" (UniqueName: \"kubernetes.io/projected/5338fe04-633c-405b-8079-1001859b804f-kube-api-access-88kvk\") pod \"frr-k8s-webhook-server-7fcb986d4-sbt5p\" (UID: \"5338fe04-633c-405b-8079-1001859b804f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.666780 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-metrics-certs\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.666885 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzvnp\" (UniqueName: \"kubernetes.io/projected/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-kube-api-access-fzvnp\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.666920 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl5hn\" (UniqueName: \"kubernetes.io/projected/fa194652-baa4-424b-bcd0-736ac7295ec4-kube-api-access-zl5hn\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.666945 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-metrics-certs\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.666976 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.667003 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-cert\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.667056 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fa194652-baa4-424b-bcd0-736ac7295ec4-metallb-excludel2\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: E1206 05:56:17.681985 4957 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 06 05:56:17 crc kubenswrapper[4957]: E1206 05:56:17.682304 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-metrics-certs podName:5ee6dee4-31cd-4d22-905c-a0339c07f1b1 nodeName:}" failed. No retries permitted until 2025-12-06 05:56:18.182281235 +0000 UTC m=+1017.832548867 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-metrics-certs") pod "controller-f8648f98b-qzfkk" (UID: "5ee6dee4-31cd-4d22-905c-a0339c07f1b1") : secret "controller-certs-secret" not found Dec 06 05:56:17 crc kubenswrapper[4957]: E1206 05:56:17.682920 4957 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 06 05:56:17 crc kubenswrapper[4957]: E1206 05:56:17.683130 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-metrics-certs podName:fa194652-baa4-424b-bcd0-736ac7295ec4 nodeName:}" failed. No retries permitted until 2025-12-06 05:56:18.183117218 +0000 UTC m=+1017.833384860 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-metrics-certs") pod "speaker-2jbdq" (UID: "fa194652-baa4-424b-bcd0-736ac7295ec4") : secret "speaker-certs-secret" not found Dec 06 05:56:17 crc kubenswrapper[4957]: E1206 05:56:17.683264 4957 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 05:56:17 crc kubenswrapper[4957]: E1206 05:56:17.683365 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist podName:fa194652-baa4-424b-bcd0-736ac7295ec4 nodeName:}" failed. No retries permitted until 2025-12-06 05:56:18.183354265 +0000 UTC m=+1017.833621897 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist") pod "speaker-2jbdq" (UID: "fa194652-baa4-424b-bcd0-736ac7295ec4") : secret "metallb-memberlist" not found Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.667852 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fa194652-baa4-424b-bcd0-736ac7295ec4-metallb-excludel2\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.698398 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-cert\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.702574 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzvnp\" (UniqueName: \"kubernetes.io/projected/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-kube-api-access-fzvnp\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.702877 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.708086 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl5hn\" (UniqueName: \"kubernetes.io/projected/fa194652-baa4-424b-bcd0-736ac7295ec4-kube-api-access-zl5hn\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:17 crc kubenswrapper[4957]: I1206 05:56:17.714168 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.219644 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p"] Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.279025 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-metrics-certs\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.279154 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-metrics-certs\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.279200 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:18 crc kubenswrapper[4957]: E1206 05:56:18.279425 4957 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 05:56:18 crc kubenswrapper[4957]: E1206 05:56:18.279523 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist podName:fa194652-baa4-424b-bcd0-736ac7295ec4 nodeName:}" failed. No retries permitted until 2025-12-06 05:56:19.279496116 +0000 UTC m=+1018.929763758 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist") pod "speaker-2jbdq" (UID: "fa194652-baa4-424b-bcd0-736ac7295ec4") : secret "metallb-memberlist" not found Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.285014 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5ee6dee4-31cd-4d22-905c-a0339c07f1b1-metrics-certs\") pod \"controller-f8648f98b-qzfkk\" (UID: \"5ee6dee4-31cd-4d22-905c-a0339c07f1b1\") " pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.285044 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-metrics-certs\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.451176 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.650242 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-qzfkk"] Dec 06 05:56:18 crc kubenswrapper[4957]: W1206 05:56:18.654578 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ee6dee4_31cd_4d22_905c_a0339c07f1b1.slice/crio-f3c3d0900c32849f1819592eb65217399e9ec8cdba0a2e292afbf84d8861f417 WatchSource:0}: Error finding container f3c3d0900c32849f1819592eb65217399e9ec8cdba0a2e292afbf84d8861f417: Status 404 returned error can't find the container with id f3c3d0900c32849f1819592eb65217399e9ec8cdba0a2e292afbf84d8861f417 Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.760072 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qzfkk" event={"ID":"5ee6dee4-31cd-4d22-905c-a0339c07f1b1","Type":"ContainerStarted","Data":"f3c3d0900c32849f1819592eb65217399e9ec8cdba0a2e292afbf84d8861f417"} Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.761603 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerStarted","Data":"e8a4557df0964b4334a21750aa7acae52be670a6a5fc72d10b0e65a1ea06ec64"} Dec 06 05:56:18 crc kubenswrapper[4957]: I1206 05:56:18.762627 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" event={"ID":"5338fe04-633c-405b-8079-1001859b804f","Type":"ContainerStarted","Data":"12a795414fe72900c4b4c02fc9ef112f0d8db10799ec8d382e95617b0135fd1b"} Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.292342 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.303583 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fa194652-baa4-424b-bcd0-736ac7295ec4-memberlist\") pod \"speaker-2jbdq\" (UID: \"fa194652-baa4-424b-bcd0-736ac7295ec4\") " pod="metallb-system/speaker-2jbdq" Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.308760 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-2jbdq" Dec 06 05:56:19 crc kubenswrapper[4957]: W1206 05:56:19.331356 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa194652_baa4_424b_bcd0_736ac7295ec4.slice/crio-0243ee5de63a137b0a64395be4273b7f02ce43cb9e3b6e058096cf2df21898a3 WatchSource:0}: Error finding container 0243ee5de63a137b0a64395be4273b7f02ce43cb9e3b6e058096cf2df21898a3: Status 404 returned error can't find the container with id 0243ee5de63a137b0a64395be4273b7f02ce43cb9e3b6e058096cf2df21898a3 Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.777971 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qzfkk" event={"ID":"5ee6dee4-31cd-4d22-905c-a0339c07f1b1","Type":"ContainerStarted","Data":"e2e310513ff3aeb379339d58e8afdaea99c242bec6c985e67483e2e6b69e2054"} Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.778187 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qzfkk" event={"ID":"5ee6dee4-31cd-4d22-905c-a0339c07f1b1","Type":"ContainerStarted","Data":"d251734b969fc640bfdeb5c7510d965272f0fe43e4882f63bb54645a0e416e44"} Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.778501 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.781798 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2jbdq" event={"ID":"fa194652-baa4-424b-bcd0-736ac7295ec4","Type":"ContainerStarted","Data":"7e04f13db6bcf46142c1ac22f1f04bb3abd7a16521544351d11971fd11804b48"} Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.781821 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2jbdq" event={"ID":"fa194652-baa4-424b-bcd0-736ac7295ec4","Type":"ContainerStarted","Data":"0243ee5de63a137b0a64395be4273b7f02ce43cb9e3b6e058096cf2df21898a3"} Dec 06 05:56:19 crc kubenswrapper[4957]: I1206 05:56:19.812321 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-qzfkk" podStartSLOduration=2.812278768 podStartE2EDuration="2.812278768s" podCreationTimestamp="2025-12-06 05:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:56:19.808496262 +0000 UTC m=+1019.458763914" watchObservedRunningTime="2025-12-06 05:56:19.812278768 +0000 UTC m=+1019.462546400" Dec 06 05:56:20 crc kubenswrapper[4957]: I1206 05:56:20.791237 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-2jbdq" event={"ID":"fa194652-baa4-424b-bcd0-736ac7295ec4","Type":"ContainerStarted","Data":"aa60455eda6bf940961ea2887c42449fae9da283223bb1fc3088c27510d2ac7a"} Dec 06 05:56:20 crc kubenswrapper[4957]: I1206 05:56:20.791294 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-2jbdq" Dec 06 05:56:20 crc kubenswrapper[4957]: I1206 05:56:20.815382 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-2jbdq" podStartSLOduration=3.815361958 podStartE2EDuration="3.815361958s" podCreationTimestamp="2025-12-06 05:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:56:20.815312036 +0000 UTC m=+1020.465579688" watchObservedRunningTime="2025-12-06 05:56:20.815361958 +0000 UTC m=+1020.465629600" Dec 06 05:56:28 crc kubenswrapper[4957]: I1206 05:56:28.455033 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-qzfkk" Dec 06 05:56:29 crc kubenswrapper[4957]: I1206 05:56:29.311460 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-2jbdq" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.022702 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-bv5w7"] Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.023480 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bv5w7" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.027815 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-c9ggd" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.027884 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.033676 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.037147 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bv5w7"] Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.108087 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkmbb\" (UniqueName: \"kubernetes.io/projected/5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e-kube-api-access-jkmbb\") pod \"openstack-operator-index-bv5w7\" (UID: \"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e\") " pod="openstack-operators/openstack-operator-index-bv5w7" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.213934 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkmbb\" (UniqueName: \"kubernetes.io/projected/5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e-kube-api-access-jkmbb\") pod \"openstack-operator-index-bv5w7\" (UID: \"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e\") " pod="openstack-operators/openstack-operator-index-bv5w7" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.249737 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkmbb\" (UniqueName: \"kubernetes.io/projected/5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e-kube-api-access-jkmbb\") pod \"openstack-operator-index-bv5w7\" (UID: \"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e\") " pod="openstack-operators/openstack-operator-index-bv5w7" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.367981 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bv5w7" Dec 06 05:56:32 crc kubenswrapper[4957]: I1206 05:56:32.572184 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bv5w7"] Dec 06 05:56:33 crc kubenswrapper[4957]: I1206 05:56:33.021233 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bv5w7" event={"ID":"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e","Type":"ContainerStarted","Data":"17a329d176a26e73e0558cc41bccd93aeed6454b2caf5231d206cfe496511aa3"} Dec 06 05:56:35 crc kubenswrapper[4957]: I1206 05:56:35.210622 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bv5w7"] Dec 06 05:56:35 crc kubenswrapper[4957]: I1206 05:56:35.819421 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-dr5bc"] Dec 06 05:56:35 crc kubenswrapper[4957]: I1206 05:56:35.822625 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:56:35 crc kubenswrapper[4957]: I1206 05:56:35.832431 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dr5bc"] Dec 06 05:56:35 crc kubenswrapper[4957]: I1206 05:56:35.974889 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8fq6\" (UniqueName: \"kubernetes.io/projected/931598ec-e32e-4be7-81cf-1a5993ad1226-kube-api-access-h8fq6\") pod \"openstack-operator-index-dr5bc\" (UID: \"931598ec-e32e-4be7-81cf-1a5993ad1226\") " pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:56:36 crc kubenswrapper[4957]: I1206 05:56:36.076621 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8fq6\" (UniqueName: \"kubernetes.io/projected/931598ec-e32e-4be7-81cf-1a5993ad1226-kube-api-access-h8fq6\") pod \"openstack-operator-index-dr5bc\" (UID: \"931598ec-e32e-4be7-81cf-1a5993ad1226\") " pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:56:36 crc kubenswrapper[4957]: I1206 05:56:36.093781 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8fq6\" (UniqueName: \"kubernetes.io/projected/931598ec-e32e-4be7-81cf-1a5993ad1226-kube-api-access-h8fq6\") pod \"openstack-operator-index-dr5bc\" (UID: \"931598ec-e32e-4be7-81cf-1a5993ad1226\") " pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:56:36 crc kubenswrapper[4957]: I1206 05:56:36.153579 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:56:37 crc kubenswrapper[4957]: E1206 05:56:37.025995 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a" Dec 06 05:56:37 crc kubenswrapper[4957]: E1206 05:56:37.026488 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:cp-frr-files,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a,Command:[/bin/sh -c cp -rLf /tmp/frr/* /etc/frr/],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:frr-startup,ReadOnly:false,MountPath:/tmp/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:frr-conf,ReadOnly:false,MountPath:/etc/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jr29w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*100,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*101,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-mv448_metallb-system(33816fc1-892d-45d9-b755-803b1bf561bf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 05:56:37 crc kubenswrapper[4957]: E1206 05:56:37.027699 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-mv448" podUID="33816fc1-892d-45d9-b755-803b1bf561bf" Dec 06 05:56:41 crc kubenswrapper[4957]: I1206 05:56:41.329668 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dr5bc"] Dec 06 05:56:42 crc kubenswrapper[4957]: I1206 05:56:42.081552 4957 generic.go:334] "Generic (PLEG): container finished" podID="33816fc1-892d-45d9-b755-803b1bf561bf" containerID="8c548d4cb51edb5841946f203ed745235412bbd0644bcd360806bcd16b06e088" exitCode=0 Dec 06 05:56:42 crc kubenswrapper[4957]: I1206 05:56:42.081627 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerDied","Data":"8c548d4cb51edb5841946f203ed745235412bbd0644bcd360806bcd16b06e088"} Dec 06 05:56:42 crc kubenswrapper[4957]: I1206 05:56:42.083039 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dr5bc" event={"ID":"931598ec-e32e-4be7-81cf-1a5993ad1226","Type":"ContainerStarted","Data":"a65446f09bcafa309f1f89f5b8fb0cacc6ca4d38a670120b56c8fd459722c176"} Dec 06 05:56:42 crc kubenswrapper[4957]: I1206 05:56:42.084522 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" event={"ID":"5338fe04-633c-405b-8079-1001859b804f","Type":"ContainerStarted","Data":"4d99cfb65e4ce4ed0283411d48eb8c4f210769e9fafbabc7401ee1f6a3822a0a"} Dec 06 05:56:42 crc kubenswrapper[4957]: I1206 05:56:42.084856 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:42 crc kubenswrapper[4957]: I1206 05:56:42.132430 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" podStartSLOduration=2.437809596 podStartE2EDuration="25.132415531s" podCreationTimestamp="2025-12-06 05:56:17 +0000 UTC" firstStartedPulling="2025-12-06 05:56:18.22876987 +0000 UTC m=+1017.879037502" lastFinishedPulling="2025-12-06 05:56:40.923375805 +0000 UTC m=+1040.573643437" observedRunningTime="2025-12-06 05:56:42.12835268 +0000 UTC m=+1041.778620362" watchObservedRunningTime="2025-12-06 05:56:42.132415531 +0000 UTC m=+1041.782683163" Dec 06 05:56:43 crc kubenswrapper[4957]: I1206 05:56:43.092814 4957 generic.go:334] "Generic (PLEG): container finished" podID="33816fc1-892d-45d9-b755-803b1bf561bf" containerID="3c324f4473ef57a49f2dd7582e8bf049486aa9be81c863928b73f56c697adfef" exitCode=0 Dec 06 05:56:43 crc kubenswrapper[4957]: I1206 05:56:43.092916 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerDied","Data":"3c324f4473ef57a49f2dd7582e8bf049486aa9be81c863928b73f56c697adfef"} Dec 06 05:56:44 crc kubenswrapper[4957]: I1206 05:56:44.103376 4957 generic.go:334] "Generic (PLEG): container finished" podID="33816fc1-892d-45d9-b755-803b1bf561bf" containerID="3e438692c16bc95d33c1cb3de70a43169f9c329883df6da8c4762fce77c40ed4" exitCode=0 Dec 06 05:56:44 crc kubenswrapper[4957]: I1206 05:56:44.103443 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerDied","Data":"3e438692c16bc95d33c1cb3de70a43169f9c329883df6da8c4762fce77c40ed4"} Dec 06 05:56:45 crc kubenswrapper[4957]: I1206 05:56:45.115274 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerStarted","Data":"76a64b479fabbe4e1ec59c929201261eb1ca5c122b9b2962a2ccda5c3c17feb8"} Dec 06 05:56:45 crc kubenswrapper[4957]: I1206 05:56:45.115992 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerStarted","Data":"2b3d9f460464a0c95f53796a68bec45863a6ae59d634a4984c80f35b58bbd4cf"} Dec 06 05:56:46 crc kubenswrapper[4957]: I1206 05:56:46.126061 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerStarted","Data":"0a761006e3d5833c1ad64158dd6e56f0f3b15d29ad1ecc094d62c1be62092db3"} Dec 06 05:56:46 crc kubenswrapper[4957]: I1206 05:56:46.126440 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerStarted","Data":"86866063840b6ebbcb74b277a7333a62b71052a73f52a789d20e47a9872d7a5a"} Dec 06 05:56:48 crc kubenswrapper[4957]: I1206 05:56:48.142524 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerStarted","Data":"7c53b802d9bd2517cbdd27567683ad431fdde44266cd36d748faa24c4080124c"} Dec 06 05:56:49 crc kubenswrapper[4957]: I1206 05:56:49.154224 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mv448" event={"ID":"33816fc1-892d-45d9-b755-803b1bf561bf","Type":"ContainerStarted","Data":"c834b5c63baa80ee3fc67fad026ce4784212e8197355955718e97634bb3177b9"} Dec 06 05:56:49 crc kubenswrapper[4957]: I1206 05:56:49.154383 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:49 crc kubenswrapper[4957]: I1206 05:56:49.178701 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-mv448" podStartSLOduration=-9223372004.676105 podStartE2EDuration="32.178669905s" podCreationTimestamp="2025-12-06 05:56:17 +0000 UTC" firstStartedPulling="2025-12-06 05:56:18.655672163 +0000 UTC m=+1018.305939795" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:56:49.173676309 +0000 UTC m=+1048.823944001" watchObservedRunningTime="2025-12-06 05:56:49.178669905 +0000 UTC m=+1048.828937537" Dec 06 05:56:52 crc kubenswrapper[4957]: I1206 05:56:52.715242 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:52 crc kubenswrapper[4957]: I1206 05:56:52.770737 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-mv448" Dec 06 05:56:57 crc kubenswrapper[4957]: I1206 05:56:57.708062 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sbt5p" Dec 06 05:56:57 crc kubenswrapper[4957]: I1206 05:56:57.763297 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-mv448" Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.260351 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dr5bc" event={"ID":"931598ec-e32e-4be7-81cf-1a5993ad1226","Type":"ContainerStarted","Data":"7cfd74048424acd523c26f9311ae3e60907ccf5771a1e8ef345ee2e790790002"} Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.261905 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bv5w7" event={"ID":"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e","Type":"ContainerStarted","Data":"0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79"} Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.262073 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-bv5w7" podUID="5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e" containerName="registry-server" containerID="cri-o://0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79" gracePeriod=2 Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.278272 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-dr5bc" podStartSLOduration=6.5905993590000005 podStartE2EDuration="30.278255789s" podCreationTimestamp="2025-12-06 05:56:35 +0000 UTC" firstStartedPulling="2025-12-06 05:56:41.338438114 +0000 UTC m=+1040.988705746" lastFinishedPulling="2025-12-06 05:57:05.026094534 +0000 UTC m=+1064.676362176" observedRunningTime="2025-12-06 05:57:05.273771547 +0000 UTC m=+1064.924039179" watchObservedRunningTime="2025-12-06 05:57:05.278255789 +0000 UTC m=+1064.928523421" Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.293118 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-bv5w7" podStartSLOduration=0.854062361 podStartE2EDuration="33.293099755s" podCreationTimestamp="2025-12-06 05:56:32 +0000 UTC" firstStartedPulling="2025-12-06 05:56:32.576244225 +0000 UTC m=+1032.226511857" lastFinishedPulling="2025-12-06 05:57:05.015281619 +0000 UTC m=+1064.665549251" observedRunningTime="2025-12-06 05:57:05.288127409 +0000 UTC m=+1064.938395051" watchObservedRunningTime="2025-12-06 05:57:05.293099755 +0000 UTC m=+1064.943367397" Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.609601 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bv5w7" Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.737965 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkmbb\" (UniqueName: \"kubernetes.io/projected/5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e-kube-api-access-jkmbb\") pod \"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e\" (UID: \"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e\") " Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.742777 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e-kube-api-access-jkmbb" (OuterVolumeSpecName: "kube-api-access-jkmbb") pod "5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e" (UID: "5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e"). InnerVolumeSpecName "kube-api-access-jkmbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:05 crc kubenswrapper[4957]: I1206 05:57:05.840892 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkmbb\" (UniqueName: \"kubernetes.io/projected/5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e-kube-api-access-jkmbb\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.153900 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.153950 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.189772 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.276229 4957 generic.go:334] "Generic (PLEG): container finished" podID="5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e" containerID="0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79" exitCode=0 Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.276287 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bv5w7" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.276346 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bv5w7" event={"ID":"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e","Type":"ContainerDied","Data":"0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79"} Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.276377 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bv5w7" event={"ID":"5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e","Type":"ContainerDied","Data":"17a329d176a26e73e0558cc41bccd93aeed6454b2caf5231d206cfe496511aa3"} Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.276397 4957 scope.go:117] "RemoveContainer" containerID="0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.292614 4957 scope.go:117] "RemoveContainer" containerID="0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79" Dec 06 05:57:06 crc kubenswrapper[4957]: E1206 05:57:06.293039 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79\": container with ID starting with 0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79 not found: ID does not exist" containerID="0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.293095 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79"} err="failed to get container status \"0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79\": rpc error: code = NotFound desc = could not find container \"0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79\": container with ID starting with 0033a1dbe33a7a5d6bf3a33e9477ce7606f6c41abdea65201c73e5e3a5d6ea79 not found: ID does not exist" Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.309114 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-bv5w7"] Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.314552 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-bv5w7"] Dec 06 05:57:06 crc kubenswrapper[4957]: I1206 05:57:06.680920 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e" path="/var/lib/kubelet/pods/5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e/volumes" Dec 06 05:57:16 crc kubenswrapper[4957]: I1206 05:57:16.182739 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-dr5bc" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.207559 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45"] Dec 06 05:57:22 crc kubenswrapper[4957]: E1206 05:57:22.209466 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e" containerName="registry-server" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.209524 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e" containerName="registry-server" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.209699 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd3b9c3-a8cf-49ba-8f53-a6a332f0397e" containerName="registry-server" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.211116 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.213749 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hjl9g" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.228195 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45"] Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.291810 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk999\" (UniqueName: \"kubernetes.io/projected/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-kube-api-access-dk999\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.291975 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-util\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.292036 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-bundle\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.393368 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk999\" (UniqueName: \"kubernetes.io/projected/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-kube-api-access-dk999\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.393464 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-util\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.393489 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-bundle\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.394001 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-bundle\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.394187 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-util\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.414292 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk999\" (UniqueName: \"kubernetes.io/projected/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-kube-api-access-dk999\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.530969 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:22 crc kubenswrapper[4957]: I1206 05:57:22.736871 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45"] Dec 06 05:57:23 crc kubenswrapper[4957]: I1206 05:57:23.393103 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" event={"ID":"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6","Type":"ContainerDied","Data":"89f66e99eb1b885d48cb7ce200167b173c5e79ee0e957bbf97573a5036d0605a"} Dec 06 05:57:23 crc kubenswrapper[4957]: I1206 05:57:23.392902 4957 generic.go:334] "Generic (PLEG): container finished" podID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerID="89f66e99eb1b885d48cb7ce200167b173c5e79ee0e957bbf97573a5036d0605a" exitCode=0 Dec 06 05:57:23 crc kubenswrapper[4957]: I1206 05:57:23.393899 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" event={"ID":"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6","Type":"ContainerStarted","Data":"e05ac26bea2b84dd9366ab5875dff32e75c14ad70a78dab3061d898eec457de0"} Dec 06 05:57:23 crc kubenswrapper[4957]: I1206 05:57:23.395785 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 05:57:28 crc kubenswrapper[4957]: I1206 05:57:28.432957 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" event={"ID":"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6","Type":"ContainerStarted","Data":"12d968cedf70aae23b66339cc853c203b49817ad10b0a5146697cb7c5fa3e83d"} Dec 06 05:57:30 crc kubenswrapper[4957]: I1206 05:57:30.445680 4957 generic.go:334] "Generic (PLEG): container finished" podID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerID="12d968cedf70aae23b66339cc853c203b49817ad10b0a5146697cb7c5fa3e83d" exitCode=0 Dec 06 05:57:30 crc kubenswrapper[4957]: I1206 05:57:30.445712 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" event={"ID":"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6","Type":"ContainerDied","Data":"12d968cedf70aae23b66339cc853c203b49817ad10b0a5146697cb7c5fa3e83d"} Dec 06 05:57:31 crc kubenswrapper[4957]: I1206 05:57:31.458988 4957 generic.go:334] "Generic (PLEG): container finished" podID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerID="4c78465edd48c7552678cae22e7471d5c1b5def55cd32cb0aa35d0306becac79" exitCode=0 Dec 06 05:57:31 crc kubenswrapper[4957]: I1206 05:57:31.460138 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" event={"ID":"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6","Type":"ContainerDied","Data":"4c78465edd48c7552678cae22e7471d5c1b5def55cd32cb0aa35d0306becac79"} Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.695452 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.742671 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-util\") pod \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.743023 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk999\" (UniqueName: \"kubernetes.io/projected/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-kube-api-access-dk999\") pod \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.743099 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-bundle\") pod \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\" (UID: \"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6\") " Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.743798 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-bundle" (OuterVolumeSpecName: "bundle") pod "9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" (UID: "9b230128-8dcd-4f69-9bcc-e57a23bd3ef6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.752361 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-kube-api-access-dk999" (OuterVolumeSpecName: "kube-api-access-dk999") pod "9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" (UID: "9b230128-8dcd-4f69-9bcc-e57a23bd3ef6"). InnerVolumeSpecName "kube-api-access-dk999". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.754733 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-util" (OuterVolumeSpecName: "util") pod "9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" (UID: "9b230128-8dcd-4f69-9bcc-e57a23bd3ef6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.845093 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk999\" (UniqueName: \"kubernetes.io/projected/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-kube-api-access-dk999\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.845151 4957 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:32 crc kubenswrapper[4957]: I1206 05:57:32.845169 4957 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b230128-8dcd-4f69-9bcc-e57a23bd3ef6-util\") on node \"crc\" DevicePath \"\"" Dec 06 05:57:33 crc kubenswrapper[4957]: I1206 05:57:33.476371 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" event={"ID":"9b230128-8dcd-4f69-9bcc-e57a23bd3ef6","Type":"ContainerDied","Data":"e05ac26bea2b84dd9366ab5875dff32e75c14ad70a78dab3061d898eec457de0"} Dec 06 05:57:33 crc kubenswrapper[4957]: I1206 05:57:33.476413 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45" Dec 06 05:57:33 crc kubenswrapper[4957]: I1206 05:57:33.476428 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e05ac26bea2b84dd9366ab5875dff32e75c14ad70a78dab3061d898eec457de0" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.583715 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7"] Dec 06 05:57:38 crc kubenswrapper[4957]: E1206 05:57:38.584595 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerName="util" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.584612 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerName="util" Dec 06 05:57:38 crc kubenswrapper[4957]: E1206 05:57:38.584626 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerName="pull" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.584634 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerName="pull" Dec 06 05:57:38 crc kubenswrapper[4957]: E1206 05:57:38.584644 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerName="extract" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.584654 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerName="extract" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.584801 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b230128-8dcd-4f69-9bcc-e57a23bd3ef6" containerName="extract" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.585291 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.589193 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qcvvc" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.610746 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7"] Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.618824 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plxrr\" (UniqueName: \"kubernetes.io/projected/87dfe427-cd07-464f-9d50-ca308f553d91-kube-api-access-plxrr\") pod \"openstack-operator-controller-operator-854f9c554b-shlk7\" (UID: \"87dfe427-cd07-464f-9d50-ca308f553d91\") " pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.720358 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plxrr\" (UniqueName: \"kubernetes.io/projected/87dfe427-cd07-464f-9d50-ca308f553d91-kube-api-access-plxrr\") pod \"openstack-operator-controller-operator-854f9c554b-shlk7\" (UID: \"87dfe427-cd07-464f-9d50-ca308f553d91\") " pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.742999 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plxrr\" (UniqueName: \"kubernetes.io/projected/87dfe427-cd07-464f-9d50-ca308f553d91-kube-api-access-plxrr\") pod \"openstack-operator-controller-operator-854f9c554b-shlk7\" (UID: \"87dfe427-cd07-464f-9d50-ca308f553d91\") " pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" Dec 06 05:57:38 crc kubenswrapper[4957]: I1206 05:57:38.903696 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" Dec 06 05:57:39 crc kubenswrapper[4957]: I1206 05:57:39.326473 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7"] Dec 06 05:57:39 crc kubenswrapper[4957]: I1206 05:57:39.511998 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" event={"ID":"87dfe427-cd07-464f-9d50-ca308f553d91","Type":"ContainerStarted","Data":"dc89ef9e2e1cf71af446c79cd2ccce9cbe7717b59c27b3b9b30012f3176d138d"} Dec 06 05:57:46 crc kubenswrapper[4957]: I1206 05:57:46.561309 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" event={"ID":"87dfe427-cd07-464f-9d50-ca308f553d91","Type":"ContainerStarted","Data":"c17420074412ec86ca8d1de35c608402ea3de4235ebf31d705ac72d5220a2eb9"} Dec 06 05:57:46 crc kubenswrapper[4957]: I1206 05:57:46.562334 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" Dec 06 05:57:46 crc kubenswrapper[4957]: I1206 05:57:46.593625 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" podStartSLOduration=2.5087603019999998 podStartE2EDuration="8.593611671s" podCreationTimestamp="2025-12-06 05:57:38 +0000 UTC" firstStartedPulling="2025-12-06 05:57:39.334870984 +0000 UTC m=+1098.985138606" lastFinishedPulling="2025-12-06 05:57:45.419722343 +0000 UTC m=+1105.069989975" observedRunningTime="2025-12-06 05:57:46.592527472 +0000 UTC m=+1106.242795134" watchObservedRunningTime="2025-12-06 05:57:46.593611671 +0000 UTC m=+1106.243879303" Dec 06 05:57:58 crc kubenswrapper[4957]: I1206 05:57:58.906797 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-854f9c554b-shlk7" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.515247 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.517085 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.519285 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-b4cfp" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.522723 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.523986 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.526046 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-rlhjw" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.529109 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.537768 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.546418 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.547596 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.553427 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.555372 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-kt5ck" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.583327 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.584655 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.587312 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-lpnhr" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.599922 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.601064 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.614494 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rfxgq" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.618240 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.682017 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.694133 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.699146 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.707039 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq4nj\" (UniqueName: \"kubernetes.io/projected/c5d6f879-3842-4481-88bd-bb090eecd9fc-kube-api-access-bq4nj\") pod \"glance-operator-controller-manager-77987cd8cd-d7r5m\" (UID: \"c5d6f879-3842-4481-88bd-bb090eecd9fc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.707113 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dg42\" (UniqueName: \"kubernetes.io/projected/ce46ec44-fc83-46e0-9f85-159e05452f0f-kube-api-access-2dg42\") pod \"heat-operator-controller-manager-5f64f6f8bb-rs7b8\" (UID: \"ce46ec44-fc83-46e0-9f85-159e05452f0f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.707137 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4zc8\" (UniqueName: \"kubernetes.io/projected/bf2466dc-9cd3-4be7-9d17-846eccda13a8-kube-api-access-j4zc8\") pod \"cinder-operator-controller-manager-859b6ccc6-mzk64\" (UID: \"bf2466dc-9cd3-4be7-9d17-846eccda13a8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.707196 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfdsm\" (UniqueName: \"kubernetes.io/projected/6276ae06-7d72-409a-81f2-70ba70bf3d86-kube-api-access-gfdsm\") pod \"barbican-operator-controller-manager-7d9dfd778-rpf52\" (UID: \"6276ae06-7d72-409a-81f2-70ba70bf3d86\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.707228 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4h5f\" (UniqueName: \"kubernetes.io/projected/e7ea0507-7466-4c88-a20b-2c936c6cdccb-kube-api-access-z4h5f\") pod \"designate-operator-controller-manager-78b4bc895b-9m6rj\" (UID: \"e7ea0507-7466-4c88-a20b-2c936c6cdccb\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.709683 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-bh4jx" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.715978 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.718206 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.724421 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.724534 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nfr4w" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.725094 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.742394 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.748190 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.749597 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.752441 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-7jc9g" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.756086 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.785943 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.787326 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.789164 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.790978 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-856kc" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808607 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4h5f\" (UniqueName: \"kubernetes.io/projected/e7ea0507-7466-4c88-a20b-2c936c6cdccb-kube-api-access-z4h5f\") pod \"designate-operator-controller-manager-78b4bc895b-9m6rj\" (UID: \"e7ea0507-7466-4c88-a20b-2c936c6cdccb\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808688 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kt6x\" (UniqueName: \"kubernetes.io/projected/5904f94f-274a-4b7f-92be-5b8acc5736ab-kube-api-access-6kt6x\") pod \"horizon-operator-controller-manager-68c6d99b8f-88lgf\" (UID: \"5904f94f-274a-4b7f-92be-5b8acc5736ab\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808746 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq4nj\" (UniqueName: \"kubernetes.io/projected/c5d6f879-3842-4481-88bd-bb090eecd9fc-kube-api-access-bq4nj\") pod \"glance-operator-controller-manager-77987cd8cd-d7r5m\" (UID: \"c5d6f879-3842-4481-88bd-bb090eecd9fc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808775 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dg42\" (UniqueName: \"kubernetes.io/projected/ce46ec44-fc83-46e0-9f85-159e05452f0f-kube-api-access-2dg42\") pod \"heat-operator-controller-manager-5f64f6f8bb-rs7b8\" (UID: \"ce46ec44-fc83-46e0-9f85-159e05452f0f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808796 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4zc8\" (UniqueName: \"kubernetes.io/projected/bf2466dc-9cd3-4be7-9d17-846eccda13a8-kube-api-access-j4zc8\") pod \"cinder-operator-controller-manager-859b6ccc6-mzk64\" (UID: \"bf2466dc-9cd3-4be7-9d17-846eccda13a8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808825 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99wh9\" (UniqueName: \"kubernetes.io/projected/e956383f-8ae1-444a-b876-a85b2bcd1bcb-kube-api-access-99wh9\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808884 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.808936 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfdsm\" (UniqueName: \"kubernetes.io/projected/6276ae06-7d72-409a-81f2-70ba70bf3d86-kube-api-access-gfdsm\") pod \"barbican-operator-controller-manager-7d9dfd778-rpf52\" (UID: \"6276ae06-7d72-409a-81f2-70ba70bf3d86\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.811163 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.812407 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.819401 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-qk4qq" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.838792 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.854826 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.855938 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.856265 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfdsm\" (UniqueName: \"kubernetes.io/projected/6276ae06-7d72-409a-81f2-70ba70bf3d86-kube-api-access-gfdsm\") pod \"barbican-operator-controller-manager-7d9dfd778-rpf52\" (UID: \"6276ae06-7d72-409a-81f2-70ba70bf3d86\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.869177 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-8dp52" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.870346 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4h5f\" (UniqueName: \"kubernetes.io/projected/e7ea0507-7466-4c88-a20b-2c936c6cdccb-kube-api-access-z4h5f\") pod \"designate-operator-controller-manager-78b4bc895b-9m6rj\" (UID: \"e7ea0507-7466-4c88-a20b-2c936c6cdccb\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.872179 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.873597 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq4nj\" (UniqueName: \"kubernetes.io/projected/c5d6f879-3842-4481-88bd-bb090eecd9fc-kube-api-access-bq4nj\") pod \"glance-operator-controller-manager-77987cd8cd-d7r5m\" (UID: \"c5d6f879-3842-4481-88bd-bb090eecd9fc\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.877600 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4zc8\" (UniqueName: \"kubernetes.io/projected/bf2466dc-9cd3-4be7-9d17-846eccda13a8-kube-api-access-j4zc8\") pod \"cinder-operator-controller-manager-859b6ccc6-mzk64\" (UID: \"bf2466dc-9cd3-4be7-9d17-846eccda13a8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.884923 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.886428 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.889715 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dg42\" (UniqueName: \"kubernetes.io/projected/ce46ec44-fc83-46e0-9f85-159e05452f0f-kube-api-access-2dg42\") pod \"heat-operator-controller-manager-5f64f6f8bb-rs7b8\" (UID: \"ce46ec44-fc83-46e0-9f85-159e05452f0f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.896405 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qb7k2" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.907300 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.911538 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.913074 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.914455 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kt6x\" (UniqueName: \"kubernetes.io/projected/5904f94f-274a-4b7f-92be-5b8acc5736ab-kube-api-access-6kt6x\") pod \"horizon-operator-controller-manager-68c6d99b8f-88lgf\" (UID: \"5904f94f-274a-4b7f-92be-5b8acc5736ab\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.914535 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vcgw\" (UniqueName: \"kubernetes.io/projected/4affc2fc-c552-4a1b-b59e-31c7ff2cbb87-kube-api-access-4vcgw\") pod \"ironic-operator-controller-manager-6c548fd776-7nwnc\" (UID: \"4affc2fc-c552-4a1b-b59e-31c7ff2cbb87\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.914580 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99wh9\" (UniqueName: \"kubernetes.io/projected/e956383f-8ae1-444a-b876-a85b2bcd1bcb-kube-api-access-99wh9\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.914611 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.914666 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kxj4\" (UniqueName: \"kubernetes.io/projected/7aa1c48d-537b-4674-a284-1d929fc9cd16-kube-api-access-5kxj4\") pod \"keystone-operator-controller-manager-7765d96ddf-w7wgp\" (UID: \"7aa1c48d-537b-4674-a284-1d929fc9cd16\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.914690 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8lz8\" (UniqueName: \"kubernetes.io/projected/b87d0440-b095-4540-a0a4-cb71461be4bd-kube-api-access-t8lz8\") pod \"manila-operator-controller-manager-7c79b5df47-8mtps\" (UID: \"b87d0440-b095-4540-a0a4-cb71461be4bd\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.915391 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-wvr28" Dec 06 05:58:24 crc kubenswrapper[4957]: E1206 05:58:24.915522 4957 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:24 crc kubenswrapper[4957]: E1206 05:58:24.915588 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert podName:e956383f-8ae1-444a-b876-a85b2bcd1bcb nodeName:}" failed. No retries permitted until 2025-12-06 05:58:25.415563835 +0000 UTC m=+1145.065831657 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert") pod "infra-operator-controller-manager-9bc586d8b-5f5p9" (UID: "e956383f-8ae1-444a-b876-a85b2bcd1bcb") : secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.940372 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.941798 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.942809 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99wh9\" (UniqueName: \"kubernetes.io/projected/e956383f-8ae1-444a-b876-a85b2bcd1bcb-kube-api-access-99wh9\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.950508 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.959525 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-xwrtx" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.966598 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kt6x\" (UniqueName: \"kubernetes.io/projected/5904f94f-274a-4b7f-92be-5b8acc5736ab-kube-api-access-6kt6x\") pod \"horizon-operator-controller-manager-68c6d99b8f-88lgf\" (UID: \"5904f94f-274a-4b7f-92be-5b8acc5736ab\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.966742 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.971188 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65"] Dec 06 05:58:24 crc kubenswrapper[4957]: I1206 05:58:24.983153 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.014184 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.018128 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kxj4\" (UniqueName: \"kubernetes.io/projected/7aa1c48d-537b-4674-a284-1d929fc9cd16-kube-api-access-5kxj4\") pod \"keystone-operator-controller-manager-7765d96ddf-w7wgp\" (UID: \"7aa1c48d-537b-4674-a284-1d929fc9cd16\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.018173 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8lz8\" (UniqueName: \"kubernetes.io/projected/b87d0440-b095-4540-a0a4-cb71461be4bd-kube-api-access-t8lz8\") pod \"manila-operator-controller-manager-7c79b5df47-8mtps\" (UID: \"b87d0440-b095-4540-a0a4-cb71461be4bd\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.018214 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd86h\" (UniqueName: \"kubernetes.io/projected/33674637-d0b9-40c9-a482-f4f18c39a80d-kube-api-access-kd86h\") pod \"nova-operator-controller-manager-697bc559fc-rrq65\" (UID: \"33674637-d0b9-40c9-a482-f4f18c39a80d\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.018246 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj5qr\" (UniqueName: \"kubernetes.io/projected/47c9992f-497a-4d57-8c37-c344ad49c199-kube-api-access-vj5qr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2x6nl\" (UID: \"47c9992f-497a-4d57-8c37-c344ad49c199\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.018270 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bzlc\" (UniqueName: \"kubernetes.io/projected/4de58666-a4e1-4cae-99d7-2088192c51db-kube-api-access-5bzlc\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2cx79\" (UID: \"4de58666-a4e1-4cae-99d7-2088192c51db\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.018302 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vcgw\" (UniqueName: \"kubernetes.io/projected/4affc2fc-c552-4a1b-b59e-31c7ff2cbb87-kube-api-access-4vcgw\") pod \"ironic-operator-controller-manager-6c548fd776-7nwnc\" (UID: \"4affc2fc-c552-4a1b-b59e-31c7ff2cbb87\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.034176 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.044113 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vcgw\" (UniqueName: \"kubernetes.io/projected/4affc2fc-c552-4a1b-b59e-31c7ff2cbb87-kube-api-access-4vcgw\") pod \"ironic-operator-controller-manager-6c548fd776-7nwnc\" (UID: \"4affc2fc-c552-4a1b-b59e-31c7ff2cbb87\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.044215 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.045908 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.048138 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8lz8\" (UniqueName: \"kubernetes.io/projected/b87d0440-b095-4540-a0a4-cb71461be4bd-kube-api-access-t8lz8\") pod \"manila-operator-controller-manager-7c79b5df47-8mtps\" (UID: \"b87d0440-b095-4540-a0a4-cb71461be4bd\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.048767 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.049424 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5hc6b" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.060247 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.072706 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.074220 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.075490 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.078182 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-zx2xh" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.087591 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.093162 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kxj4\" (UniqueName: \"kubernetes.io/projected/7aa1c48d-537b-4674-a284-1d929fc9cd16-kube-api-access-5kxj4\") pod \"keystone-operator-controller-manager-7765d96ddf-w7wgp\" (UID: \"7aa1c48d-537b-4674-a284-1d929fc9cd16\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.093795 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.319185 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.320229 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.323197 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.328463 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.328505 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.329621 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd86h\" (UniqueName: \"kubernetes.io/projected/33674637-d0b9-40c9-a482-f4f18c39a80d-kube-api-access-kd86h\") pod \"nova-operator-controller-manager-697bc559fc-rrq65\" (UID: \"33674637-d0b9-40c9-a482-f4f18c39a80d\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.329669 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj5qr\" (UniqueName: \"kubernetes.io/projected/47c9992f-497a-4d57-8c37-c344ad49c199-kube-api-access-vj5qr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2x6nl\" (UID: \"47c9992f-497a-4d57-8c37-c344ad49c199\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.329717 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bzlc\" (UniqueName: \"kubernetes.io/projected/4de58666-a4e1-4cae-99d7-2088192c51db-kube-api-access-5bzlc\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2cx79\" (UID: \"4de58666-a4e1-4cae-99d7-2088192c51db\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.329817 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc2ks\" (UniqueName: \"kubernetes.io/projected/db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5-kube-api-access-tc2ks\") pod \"octavia-operator-controller-manager-998648c74-c2t2c\" (UID: \"db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.337749 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ghl89" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.339302 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.420350 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj5qr\" (UniqueName: \"kubernetes.io/projected/47c9992f-497a-4d57-8c37-c344ad49c199-kube-api-access-vj5qr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2x6nl\" (UID: \"47c9992f-497a-4d57-8c37-c344ad49c199\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.423298 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bzlc\" (UniqueName: \"kubernetes.io/projected/4de58666-a4e1-4cae-99d7-2088192c51db-kube-api-access-5bzlc\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2cx79\" (UID: \"4de58666-a4e1-4cae-99d7-2088192c51db\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.424536 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.432982 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.433041 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g4dl\" (UniqueName: \"kubernetes.io/projected/0a9ebb33-9805-4dec-a879-df7417c8ce7b-kube-api-access-5g4dl\") pod \"ovn-operator-controller-manager-b6456fdb6-mc7wf\" (UID: \"0a9ebb33-9805-4dec-a879-df7417c8ce7b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.433100 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fnxw\" (UniqueName: \"kubernetes.io/projected/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-kube-api-access-5fnxw\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.437365 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.437754 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvt6p\" (UniqueName: \"kubernetes.io/projected/d349e041-2168-4b8a-b6c4-a5840de8c9d8-kube-api-access-fvt6p\") pod \"placement-operator-controller-manager-78f8948974-g6zcq\" (UID: \"d349e041-2168-4b8a-b6c4-a5840de8c9d8\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.451908 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2"] Dec 06 05:58:25 crc kubenswrapper[4957]: E1206 05:58:25.443203 4957 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:25 crc kubenswrapper[4957]: E1206 05:58:25.453015 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert podName:e956383f-8ae1-444a-b876-a85b2bcd1bcb nodeName:}" failed. No retries permitted until 2025-12-06 05:58:26.452980218 +0000 UTC m=+1146.103247850 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert") pod "infra-operator-controller-manager-9bc586d8b-5f5p9" (UID: "e956383f-8ae1-444a-b876-a85b2bcd1bcb") : secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.456268 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd86h\" (UniqueName: \"kubernetes.io/projected/33674637-d0b9-40c9-a482-f4f18c39a80d-kube-api-access-kd86h\") pod \"nova-operator-controller-manager-697bc559fc-rrq65\" (UID: \"33674637-d0b9-40c9-a482-f4f18c39a80d\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.465066 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc2ks\" (UniqueName: \"kubernetes.io/projected/db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5-kube-api-access-tc2ks\") pod \"octavia-operator-controller-manager-998648c74-c2t2c\" (UID: \"db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.479508 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.496444 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-cq785" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.501096 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc2ks\" (UniqueName: \"kubernetes.io/projected/db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5-kube-api-access-tc2ks\") pod \"octavia-operator-controller-manager-998648c74-c2t2c\" (UID: \"db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.504059 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.506634 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.524995 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-cxz7x" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.537455 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.540650 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.543473 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.546842 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-c46xm" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.556898 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.566049 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fnxw\" (UniqueName: \"kubernetes.io/projected/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-kube-api-access-5fnxw\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.566177 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q84xh\" (UniqueName: \"kubernetes.io/projected/f2ab1e95-3e28-4d99-9340-f1eb24bc6dca-kube-api-access-q84xh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nr7jl\" (UID: \"f2ab1e95-3e28-4d99-9340-f1eb24bc6dca\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.566215 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvt6p\" (UniqueName: \"kubernetes.io/projected/d349e041-2168-4b8a-b6c4-a5840de8c9d8-kube-api-access-fvt6p\") pod \"placement-operator-controller-manager-78f8948974-g6zcq\" (UID: \"d349e041-2168-4b8a-b6c4-a5840de8c9d8\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.566287 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.566322 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxk6n\" (UniqueName: \"kubernetes.io/projected/2d7a7ff4-095a-4dff-9c96-25d03888a51e-kube-api-access-lxk6n\") pod \"swift-operator-controller-manager-5f8c65bbfc-xjbq2\" (UID: \"2d7a7ff4-095a-4dff-9c96-25d03888a51e\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.566344 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g4dl\" (UniqueName: \"kubernetes.io/projected/0a9ebb33-9805-4dec-a879-df7417c8ce7b-kube-api-access-5g4dl\") pod \"ovn-operator-controller-manager-b6456fdb6-mc7wf\" (UID: \"0a9ebb33-9805-4dec-a879-df7417c8ce7b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" Dec 06 05:58:25 crc kubenswrapper[4957]: E1206 05:58:25.567287 4957 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:25 crc kubenswrapper[4957]: E1206 05:58:25.567379 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert podName:f9a79638-2a84-4bc8-a3c8-3289f3a1adae nodeName:}" failed. No retries permitted until 2025-12-06 05:58:26.067351099 +0000 UTC m=+1145.717618921 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" (UID: "f9a79638-2a84-4bc8-a3c8-3289f3a1adae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.576922 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8"] Dec 06 05:58:25 crc kubenswrapper[4957]: I1206 05:58:25.578579 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.589385 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-b5q9b" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.598277 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g4dl\" (UniqueName: \"kubernetes.io/projected/0a9ebb33-9805-4dec-a879-df7417c8ce7b-kube-api-access-5g4dl\") pod \"ovn-operator-controller-manager-b6456fdb6-mc7wf\" (UID: \"0a9ebb33-9805-4dec-a879-df7417c8ce7b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.598625 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvt6p\" (UniqueName: \"kubernetes.io/projected/d349e041-2168-4b8a-b6c4-a5840de8c9d8-kube-api-access-fvt6p\") pod \"placement-operator-controller-manager-78f8948974-g6zcq\" (UID: \"d349e041-2168-4b8a-b6c4-a5840de8c9d8\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.600625 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8"] Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.602592 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.604955 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fnxw\" (UniqueName: \"kubernetes.io/projected/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-kube-api-access-5fnxw\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.619885 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.637390 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.660075 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h"] Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.661019 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h"] Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.661104 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.678045 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq2m2\" (UniqueName: \"kubernetes.io/projected/ea50a02a-23dc-4fcf-bcfd-048b66ce8152-kube-api-access-cq2m2\") pod \"test-operator-controller-manager-5854674fcc-fvj4l\" (UID: \"ea50a02a-23dc-4fcf-bcfd-048b66ce8152\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.678156 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q84xh\" (UniqueName: \"kubernetes.io/projected/f2ab1e95-3e28-4d99-9340-f1eb24bc6dca-kube-api-access-q84xh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nr7jl\" (UID: \"f2ab1e95-3e28-4d99-9340-f1eb24bc6dca\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.678194 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twmvf\" (UniqueName: \"kubernetes.io/projected/ee5ec7b9-749f-4b24-826c-bebb580f7c5b-kube-api-access-twmvf\") pod \"watcher-operator-controller-manager-769dc69bc-k4mm8\" (UID: \"ee5ec7b9-749f-4b24-826c-bebb580f7c5b\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.678274 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxk6n\" (UniqueName: \"kubernetes.io/projected/2d7a7ff4-095a-4dff-9c96-25d03888a51e-kube-api-access-lxk6n\") pod \"swift-operator-controller-manager-5f8c65bbfc-xjbq2\" (UID: \"2d7a7ff4-095a-4dff-9c96-25d03888a51e\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.680255 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.708331 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn"] Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.709443 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.715432 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn"] Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.717359 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.762156 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-t4slg" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.762289 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.762386 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.762709 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-jptnn" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.775722 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q84xh\" (UniqueName: \"kubernetes.io/projected/f2ab1e95-3e28-4d99-9340-f1eb24bc6dca-kube-api-access-q84xh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-nr7jl\" (UID: \"f2ab1e95-3e28-4d99-9340-f1eb24bc6dca\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.778079 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxk6n\" (UniqueName: \"kubernetes.io/projected/2d7a7ff4-095a-4dff-9c96-25d03888a51e-kube-api-access-lxk6n\") pod \"swift-operator-controller-manager-5f8c65bbfc-xjbq2\" (UID: \"2d7a7ff4-095a-4dff-9c96-25d03888a51e\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.779465 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.779525 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq2m2\" (UniqueName: \"kubernetes.io/projected/ea50a02a-23dc-4fcf-bcfd-048b66ce8152-kube-api-access-cq2m2\") pod \"test-operator-controller-manager-5854674fcc-fvj4l\" (UID: \"ea50a02a-23dc-4fcf-bcfd-048b66ce8152\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.779603 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdgr2\" (UniqueName: \"kubernetes.io/projected/31332329-9dd4-4665-b1d0-9662ee3d1b8e-kube-api-access-hdgr2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-6gccn\" (UID: \"31332329-9dd4-4665-b1d0-9662ee3d1b8e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.779697 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twmvf\" (UniqueName: \"kubernetes.io/projected/ee5ec7b9-749f-4b24-826c-bebb580f7c5b-kube-api-access-twmvf\") pod \"watcher-operator-controller-manager-769dc69bc-k4mm8\" (UID: \"ee5ec7b9-749f-4b24-826c-bebb580f7c5b\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.779765 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.780027 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wndrk\" (UniqueName: \"kubernetes.io/projected/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-kube-api-access-wndrk\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.794058 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.809235 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twmvf\" (UniqueName: \"kubernetes.io/projected/ee5ec7b9-749f-4b24-826c-bebb580f7c5b-kube-api-access-twmvf\") pod \"watcher-operator-controller-manager-769dc69bc-k4mm8\" (UID: \"ee5ec7b9-749f-4b24-826c-bebb580f7c5b\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.812364 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq2m2\" (UniqueName: \"kubernetes.io/projected/ea50a02a-23dc-4fcf-bcfd-048b66ce8152-kube-api-access-cq2m2\") pod \"test-operator-controller-manager-5854674fcc-fvj4l\" (UID: \"ea50a02a-23dc-4fcf-bcfd-048b66ce8152\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.829849 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.882521 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdgr2\" (UniqueName: \"kubernetes.io/projected/31332329-9dd4-4665-b1d0-9662ee3d1b8e-kube-api-access-hdgr2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-6gccn\" (UID: \"31332329-9dd4-4665-b1d0-9662ee3d1b8e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.882574 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.882641 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wndrk\" (UniqueName: \"kubernetes.io/projected/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-kube-api-access-wndrk\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.882672 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:25.882802 4957 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:25.882869 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:26.382852165 +0000 UTC m=+1146.033119797 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "metrics-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:25.883013 4957 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:25.883095 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:26.383070842 +0000 UTC m=+1146.033338474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:25.965687 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wndrk\" (UniqueName: \"kubernetes.io/projected/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-kube-api-access-wndrk\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.089802 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.090086 4957 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.090335 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert podName:f9a79638-2a84-4bc8-a3c8-3289f3a1adae nodeName:}" failed. No retries permitted until 2025-12-06 05:58:27.090315271 +0000 UTC m=+1146.740582903 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" (UID: "f9a79638-2a84-4bc8-a3c8-3289f3a1adae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.149426 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.162993 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.312238 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.423118 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj"] Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.425383 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.430800 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.425494 4957 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.431146 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:27.431130352 +0000 UTC m=+1147.081397984 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "metrics-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.431092 4957 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.431806 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:27.43179737 +0000 UTC m=+1147.082065002 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:26.532493 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.532957 4957 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:26.533029 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert podName:e956383f-8ae1-444a-b876-a85b2bcd1bcb nodeName:}" failed. No retries permitted until 2025-12-06 05:58:28.533006266 +0000 UTC m=+1148.183273908 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert") pod "infra-operator-controller-manager-9bc586d8b-5f5p9" (UID: "e956383f-8ae1-444a-b876-a85b2bcd1bcb") : secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.570881 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.570978 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.571024 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.571131 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.571310 4957 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.571366 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:30.57135178 +0000 UTC m=+1150.221619412 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "metrics-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.614353 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdgr2\" (UniqueName: \"kubernetes.io/projected/31332329-9dd4-4665-b1d0-9662ee3d1b8e-kube-api-access-hdgr2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-6gccn\" (UID: \"31332329-9dd4-4665-b1d0-9662ee3d1b8e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.619551 4957 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.619704 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert podName:e956383f-8ae1-444a-b876-a85b2bcd1bcb nodeName:}" failed. No retries permitted until 2025-12-06 05:58:32.619644495 +0000 UTC m=+1152.269912137 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert") pod "infra-operator-controller-manager-9bc586d8b-5f5p9" (UID: "e956383f-8ae1-444a-b876-a85b2bcd1bcb") : secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.627252 4957 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.627363 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert podName:f9a79638-2a84-4bc8-a3c8-3289f3a1adae nodeName:}" failed. No retries permitted until 2025-12-06 05:58:30.627327843 +0000 UTC m=+1150.277595485 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" (UID: "f9a79638-2a84-4bc8-a3c8-3289f3a1adae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.627465 4957 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: E1206 05:58:28.627515 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:30.627500077 +0000 UTC m=+1150.277767729 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "webhook-server-cert" not found Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.691419 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" event={"ID":"e7ea0507-7466-4c88-a20b-2c936c6cdccb","Type":"ContainerStarted","Data":"603a2f14046cdf9f04f3a3adde3625ad2cc49f57f1aaa6cc6e64ba58c3225e10"} Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.731649 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" Dec 06 05:58:28 crc kubenswrapper[4957]: I1206 05:58:28.987493 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.000479 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.377109 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8"] Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.380887 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce46ec44_fc83_46e0_9f85_159e05452f0f.slice/crio-87c017ae4959a9121181214eccb614d53dfb17b9ff80ca8a8adafc5c827225b2 WatchSource:0}: Error finding container 87c017ae4959a9121181214eccb614d53dfb17b9ff80ca8a8adafc5c827225b2: Status 404 returned error can't find the container with id 87c017ae4959a9121181214eccb614d53dfb17b9ff80ca8a8adafc5c827225b2 Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.393817 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.398530 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.403874 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.420162 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.434044 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.440517 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.448366 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.452938 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps"] Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.458990 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6276ae06_7d72_409a_81f2_70ba70bf3d86.slice/crio-6441f0ccfe3d47be2682e8c010cbcc7c58d4496cd660d218a997f9af71d198b4 WatchSource:0}: Error finding container 6441f0ccfe3d47be2682e8c010cbcc7c58d4496cd660d218a997f9af71d198b4: Status 404 returned error can't find the container with id 6441f0ccfe3d47be2682e8c010cbcc7c58d4496cd660d218a997f9af71d198b4 Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.461550 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d7a7ff4_095a_4dff_9c96_25d03888a51e.slice/crio-99136191201e0811c120741451f77be21f1139e1234743e287e428e63466d941 WatchSource:0}: Error finding container 99136191201e0811c120741451f77be21f1139e1234743e287e428e63466d941: Status 404 returned error can't find the container with id 99136191201e0811c120741451f77be21f1139e1234743e287e428e63466d941 Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.461770 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.468008 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl"] Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.473206 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf"] Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.476167 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb87d0440_b095_4540_a0a4_cb71461be4bd.slice/crio-83be7276782f996020c96952d68e23eabb0fc11276b1e51e242c21e6c5d7fa2a WatchSource:0}: Error finding container 83be7276782f996020c96952d68e23eabb0fc11276b1e51e242c21e6c5d7fa2a: Status 404 returned error can't find the container with id 83be7276782f996020c96952d68e23eabb0fc11276b1e51e242c21e6c5d7fa2a Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.478519 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5g4dl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-mc7wf_openstack-operators(0a9ebb33-9805-4dec-a879-df7417c8ce7b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.479100 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t8lz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-8mtps_openstack-operators(b87d0440-b095-4540-a0a4-cb71461be4bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.480749 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5g4dl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-mc7wf_openstack-operators(0a9ebb33-9805-4dec-a879-df7417c8ce7b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.481204 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t8lz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-8mtps_openstack-operators(b87d0440-b095-4540-a0a4-cb71461be4bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.482033 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" podUID="0a9ebb33-9805-4dec-a879-df7417c8ce7b" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.482287 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" podUID="b87d0440-b095-4540-a0a4-cb71461be4bd" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.482436 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5kxj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-w7wgp_openstack-operators(7aa1c48d-537b-4674-a284-1d929fc9cd16): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.485283 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5kxj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-w7wgp_openstack-operators(7aa1c48d-537b-4674-a284-1d929fc9cd16): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.486417 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" podUID="7aa1c48d-537b-4674-a284-1d929fc9cd16" Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.486880 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4affc2fc_c552_4a1b_b59e_31c7ff2cbb87.slice/crio-285e078811fc86e894e738b062c38aa006d07b9a9a7ecbcd44ab9fe3de43f333 WatchSource:0}: Error finding container 285e078811fc86e894e738b062c38aa006d07b9a9a7ecbcd44ab9fe3de43f333: Status 404 returned error can't find the container with id 285e078811fc86e894e738b062c38aa006d07b9a9a7ecbcd44ab9fe3de43f333 Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.489061 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65"] Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.491678 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47c9992f_497a_4d57_8c37_c344ad49c199.slice/crio-c8dfbc0c46f94d42724cb76b4b5e60a53d1f5fbc12e94e7c7a5736dcf6f7a3dc WatchSource:0}: Error finding container c8dfbc0c46f94d42724cb76b4b5e60a53d1f5fbc12e94e7c7a5736dcf6f7a3dc: Status 404 returned error can't find the container with id c8dfbc0c46f94d42724cb76b4b5e60a53d1f5fbc12e94e7c7a5736dcf6f7a3dc Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.494059 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd349e041_2168_4b8a_b6c4_a5840de8c9d8.slice/crio-6c748b0ab518de04c6184b4c1b2aff0ca560e2656f32f2fa04e5895103500f74 WatchSource:0}: Error finding container 6c748b0ab518de04c6184b4c1b2aff0ca560e2656f32f2fa04e5895103500f74: Status 404 returned error can't find the container with id 6c748b0ab518de04c6184b4c1b2aff0ca560e2656f32f2fa04e5895103500f74 Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.494739 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp"] Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.494785 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4vcgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-7nwnc_openstack-operators(4affc2fc-c552-4a1b-b59e-31c7ff2cbb87): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.494916 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vj5qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-2x6nl_openstack-operators(47c9992f-497a-4d57-8c37-c344ad49c199): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.497255 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4vcgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-7nwnc_openstack-operators(4affc2fc-c552-4a1b-b59e-31c7ff2cbb87): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.497755 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vj5qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-2x6nl_openstack-operators(47c9992f-497a-4d57-8c37-c344ad49c199): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.498572 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" podUID="4affc2fc-c552-4a1b-b59e-31c7ff2cbb87" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.498674 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fvt6p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-g6zcq_openstack-operators(d349e041-2168-4b8a-b6c4-a5840de8c9d8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.499325 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2"] Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.499561 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" podUID="47c9992f-497a-4d57-8c37-c344ad49c199" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.500943 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fvt6p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-g6zcq_openstack-operators(d349e041-2168-4b8a-b6c4-a5840de8c9d8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.502171 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" podUID="d349e041-2168-4b8a-b6c4-a5840de8c9d8" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.503361 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq"] Dec 06 05:58:29 crc kubenswrapper[4957]: W1206 05:58:29.506845 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31332329_9dd4_4665_b1d0_9662ee3d1b8e.slice/crio-ea19cf0497ed7c58ec400ff31949fdbf2a21131968922c5a5bcaf903eda7f294 WatchSource:0}: Error finding container ea19cf0497ed7c58ec400ff31949fdbf2a21131968922c5a5bcaf903eda7f294: Status 404 returned error can't find the container with id ea19cf0497ed7c58ec400ff31949fdbf2a21131968922c5a5bcaf903eda7f294 Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.507249 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn"] Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.511355 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hdgr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-6gccn_openstack-operators(31332329-9dd4-4665-b1d0-9662ee3d1b8e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.512874 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" podUID="31332329-9dd4-4665-b1d0-9662ee3d1b8e" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.665036 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" event={"ID":"ee5ec7b9-749f-4b24-826c-bebb580f7c5b","Type":"ContainerStarted","Data":"e9d762aebc25dd27b481f8f612c7a8a8c8d96526a99a22ebdd2cbae0688e186d"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.666596 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" event={"ID":"ce46ec44-fc83-46e0-9f85-159e05452f0f","Type":"ContainerStarted","Data":"87c017ae4959a9121181214eccb614d53dfb17b9ff80ca8a8adafc5c827225b2"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.667764 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" event={"ID":"d349e041-2168-4b8a-b6c4-a5840de8c9d8","Type":"ContainerStarted","Data":"6c748b0ab518de04c6184b4c1b2aff0ca560e2656f32f2fa04e5895103500f74"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.670906 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" event={"ID":"31332329-9dd4-4665-b1d0-9662ee3d1b8e","Type":"ContainerStarted","Data":"ea19cf0497ed7c58ec400ff31949fdbf2a21131968922c5a5bcaf903eda7f294"} Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.671517 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" podUID="d349e041-2168-4b8a-b6c4-a5840de8c9d8" Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.673002 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" podUID="31332329-9dd4-4665-b1d0-9662ee3d1b8e" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.673883 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" event={"ID":"47c9992f-497a-4d57-8c37-c344ad49c199","Type":"ContainerStarted","Data":"c8dfbc0c46f94d42724cb76b4b5e60a53d1f5fbc12e94e7c7a5736dcf6f7a3dc"} Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.675937 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" podUID="47c9992f-497a-4d57-8c37-c344ad49c199" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.677104 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" event={"ID":"6276ae06-7d72-409a-81f2-70ba70bf3d86","Type":"ContainerStarted","Data":"6441f0ccfe3d47be2682e8c010cbcc7c58d4496cd660d218a997f9af71d198b4"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.679903 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" event={"ID":"5904f94f-274a-4b7f-92be-5b8acc5736ab","Type":"ContainerStarted","Data":"723ffa52cbf4c517b1db198253a30574d9be4e54bc22c9379b56d76485dab37c"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.685449 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" event={"ID":"b87d0440-b095-4540-a0a4-cb71461be4bd","Type":"ContainerStarted","Data":"83be7276782f996020c96952d68e23eabb0fc11276b1e51e242c21e6c5d7fa2a"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.691006 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" event={"ID":"c5d6f879-3842-4481-88bd-bb090eecd9fc","Type":"ContainerStarted","Data":"3192ce45069ad659bdb34b73974d591361d76bd87ff987fa5125741a57368f7f"} Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.691728 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" podUID="b87d0440-b095-4540-a0a4-cb71461be4bd" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.692401 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" event={"ID":"0a9ebb33-9805-4dec-a879-df7417c8ce7b","Type":"ContainerStarted","Data":"638e8c8ecb52d8f504a28f9f94bd30386ea33ad7c553afe55889ecb2532fc950"} Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.694642 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" podUID="0a9ebb33-9805-4dec-a879-df7417c8ce7b" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.705474 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" event={"ID":"bf2466dc-9cd3-4be7-9d17-846eccda13a8","Type":"ContainerStarted","Data":"bdb28c7dfcac04cb1ee8957476d755b578d2bdc91d5db8a0f40b6b0c6c3ae1e3"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.711746 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" event={"ID":"db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5","Type":"ContainerStarted","Data":"f45100de105fcf182ac53046f6157946a10ca0c3cc2d1374a8596ed753eca916"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.713039 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" event={"ID":"33674637-d0b9-40c9-a482-f4f18c39a80d","Type":"ContainerStarted","Data":"d6d084851dfcf6c7425decac8cbb41f4776ddfc7066ba12fab9eca9cd253b37d"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.714202 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" event={"ID":"4affc2fc-c552-4a1b-b59e-31c7ff2cbb87","Type":"ContainerStarted","Data":"285e078811fc86e894e738b062c38aa006d07b9a9a7ecbcd44ab9fe3de43f333"} Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.717277 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" podUID="4affc2fc-c552-4a1b-b59e-31c7ff2cbb87" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.717485 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" event={"ID":"f2ab1e95-3e28-4d99-9340-f1eb24bc6dca","Type":"ContainerStarted","Data":"352a77cb886450355c26763352449718274f7e2ba98d0692f1a1d36639f0d977"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.718329 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" event={"ID":"2d7a7ff4-095a-4dff-9c96-25d03888a51e","Type":"ContainerStarted","Data":"99136191201e0811c120741451f77be21f1139e1234743e287e428e63466d941"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.720955 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" event={"ID":"7aa1c48d-537b-4674-a284-1d929fc9cd16","Type":"ContainerStarted","Data":"703900b9c3d331cae10e197ce3d45b289857277f7e33f4628cef86a53084b5ff"} Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.721886 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" event={"ID":"4de58666-a4e1-4cae-99d7-2088192c51db","Type":"ContainerStarted","Data":"563850d1bc974236d941e21f693338fae801abe4ffbfdc468c53df933d119728"} Dec 06 05:58:29 crc kubenswrapper[4957]: E1206 05:58:29.722269 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" podUID="7aa1c48d-537b-4674-a284-1d929fc9cd16" Dec 06 05:58:29 crc kubenswrapper[4957]: I1206 05:58:29.722657 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" event={"ID":"ea50a02a-23dc-4fcf-bcfd-048b66ce8152","Type":"ContainerStarted","Data":"f2b97f560484e5a56ff943fc3cb42a1a7236bcf25f64283ce0e0e7db3c1f982f"} Dec 06 05:58:30 crc kubenswrapper[4957]: I1206 05:58:30.607330 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.607528 4957 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.607900 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:34.607878056 +0000 UTC m=+1154.258145698 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "metrics-server-cert" not found Dec 06 05:58:30 crc kubenswrapper[4957]: I1206 05:58:30.709854 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:30 crc kubenswrapper[4957]: I1206 05:58:30.709959 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.710106 4957 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.710197 4957 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.710203 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:34.710180042 +0000 UTC m=+1154.360447734 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "webhook-server-cert" not found Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.710297 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert podName:f9a79638-2a84-4bc8-a3c8-3289f3a1adae nodeName:}" failed. No retries permitted until 2025-12-06 05:58:34.710279874 +0000 UTC m=+1154.360547506 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" (UID: "f9a79638-2a84-4bc8-a3c8-3289f3a1adae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.735220 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" podUID="31332329-9dd4-4665-b1d0-9662ee3d1b8e" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.735658 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" podUID="4affc2fc-c552-4a1b-b59e-31c7ff2cbb87" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.735803 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" podUID="b87d0440-b095-4540-a0a4-cb71461be4bd" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.735888 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" podUID="47c9992f-497a-4d57-8c37-c344ad49c199" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.735957 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" podUID="0a9ebb33-9805-4dec-a879-df7417c8ce7b" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.736407 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" podUID="7aa1c48d-537b-4674-a284-1d929fc9cd16" Dec 06 05:58:30 crc kubenswrapper[4957]: E1206 05:58:30.737760 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" podUID="d349e041-2168-4b8a-b6c4-a5840de8c9d8" Dec 06 05:58:32 crc kubenswrapper[4957]: I1206 05:58:32.648524 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:32 crc kubenswrapper[4957]: E1206 05:58:32.648750 4957 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:32 crc kubenswrapper[4957]: E1206 05:58:32.649023 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert podName:e956383f-8ae1-444a-b876-a85b2bcd1bcb nodeName:}" failed. No retries permitted until 2025-12-06 05:58:40.649002627 +0000 UTC m=+1160.299270259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert") pod "infra-operator-controller-manager-9bc586d8b-5f5p9" (UID: "e956383f-8ae1-444a-b876-a85b2bcd1bcb") : secret "infra-operator-webhook-server-cert" not found Dec 06 05:58:34 crc kubenswrapper[4957]: I1206 05:58:34.653396 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:34 crc kubenswrapper[4957]: E1206 05:58:34.653641 4957 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 05:58:34 crc kubenswrapper[4957]: E1206 05:58:34.653734 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:42.653715143 +0000 UTC m=+1162.303982775 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "metrics-server-cert" not found Dec 06 05:58:34 crc kubenswrapper[4957]: I1206 05:58:34.754688 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:34 crc kubenswrapper[4957]: I1206 05:58:34.754812 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:34 crc kubenswrapper[4957]: E1206 05:58:34.755011 4957 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:34 crc kubenswrapper[4957]: E1206 05:58:34.755126 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert podName:f9a79638-2a84-4bc8-a3c8-3289f3a1adae nodeName:}" failed. No retries permitted until 2025-12-06 05:58:42.755094684 +0000 UTC m=+1162.405362366 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" (UID: "f9a79638-2a84-4bc8-a3c8-3289f3a1adae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:34 crc kubenswrapper[4957]: E1206 05:58:34.755012 4957 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 05:58:34 crc kubenswrapper[4957]: E1206 05:58:34.755200 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs podName:b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e nodeName:}" failed. No retries permitted until 2025-12-06 05:58:42.755182516 +0000 UTC m=+1162.405450198 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs") pod "openstack-operator-controller-manager-75c7c544f9-bnr4h" (UID: "b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e") : secret "webhook-server-cert" not found Dec 06 05:58:40 crc kubenswrapper[4957]: I1206 05:58:40.180873 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:58:40 crc kubenswrapper[4957]: I1206 05:58:40.181552 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:58:40 crc kubenswrapper[4957]: I1206 05:58:40.749211 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:40 crc kubenswrapper[4957]: I1206 05:58:40.757101 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e956383f-8ae1-444a-b876-a85b2bcd1bcb-cert\") pod \"infra-operator-controller-manager-9bc586d8b-5f5p9\" (UID: \"e956383f-8ae1-444a-b876-a85b2bcd1bcb\") " pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:40 crc kubenswrapper[4957]: I1206 05:58:40.951575 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:58:42 crc kubenswrapper[4957]: I1206 05:58:42.678287 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:42 crc kubenswrapper[4957]: I1206 05:58:42.685165 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-metrics-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:42 crc kubenswrapper[4957]: I1206 05:58:42.780495 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:42 crc kubenswrapper[4957]: E1206 05:58:42.781020 4957 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:42 crc kubenswrapper[4957]: E1206 05:58:42.781257 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert podName:f9a79638-2a84-4bc8-a3c8-3289f3a1adae nodeName:}" failed. No retries permitted until 2025-12-06 05:58:58.781212126 +0000 UTC m=+1178.431479828 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" (UID: "f9a79638-2a84-4bc8-a3c8-3289f3a1adae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 05:58:42 crc kubenswrapper[4957]: I1206 05:58:42.781373 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:42 crc kubenswrapper[4957]: I1206 05:58:42.793265 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e-webhook-certs\") pod \"openstack-operator-controller-manager-75c7c544f9-bnr4h\" (UID: \"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e\") " pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:43 crc kubenswrapper[4957]: I1206 05:58:43.061391 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:58:47 crc kubenswrapper[4957]: E1206 05:58:47.133679 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 06 05:58:47 crc kubenswrapper[4957]: E1206 05:58:47.134389 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lxk6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-xjbq2_openstack-operators(2d7a7ff4-095a-4dff-9c96-25d03888a51e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:58:49 crc kubenswrapper[4957]: E1206 05:58:49.017365 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 06 05:58:49 crc kubenswrapper[4957]: E1206 05:58:49.017888 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q84xh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-nr7jl_openstack-operators(f2ab1e95-3e28-4d99-9340-f1eb24bc6dca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:58:49 crc kubenswrapper[4957]: E1206 05:58:49.668514 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 06 05:58:49 crc kubenswrapper[4957]: E1206 05:58:49.668737 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6kt6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-88lgf_openstack-operators(5904f94f-274a-4b7f-92be-5b8acc5736ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:58:50 crc kubenswrapper[4957]: E1206 05:58:50.928577 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 06 05:58:50 crc kubenswrapper[4957]: E1206 05:58:50.929147 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2dg42,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-rs7b8_openstack-operators(ce46ec44-fc83-46e0-9f85-159e05452f0f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:58:51 crc kubenswrapper[4957]: E1206 05:58:51.537902 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 06 05:58:51 crc kubenswrapper[4957]: E1206 05:58:51.538079 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bq4nj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-d7r5m_openstack-operators(c5d6f879-3842-4481-88bd-bb090eecd9fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:58:52 crc kubenswrapper[4957]: E1206 05:58:52.020030 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 06 05:58:52 crc kubenswrapper[4957]: E1206 05:58:52.020266 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kd86h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-rrq65_openstack-operators(33674637-d0b9-40c9-a482-f4f18c39a80d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:58:58 crc kubenswrapper[4957]: I1206 05:58:58.842340 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:58 crc kubenswrapper[4957]: I1206 05:58:58.852748 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9a79638-2a84-4bc8-a3c8-3289f3a1adae-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8\" (UID: \"f9a79638-2a84-4bc8-a3c8-3289f3a1adae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:58:58 crc kubenswrapper[4957]: I1206 05:58:58.922070 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:59:00 crc kubenswrapper[4957]: I1206 05:59:00.267786 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9"] Dec 06 05:59:00 crc kubenswrapper[4957]: W1206 05:59:00.450692 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode956383f_8ae1_444a_b876_a85b2bcd1bcb.slice/crio-772bed82dcb1aa07361825b7f6893d507fcfa74c070d6f5df70ec5be90c33778 WatchSource:0}: Error finding container 772bed82dcb1aa07361825b7f6893d507fcfa74c070d6f5df70ec5be90c33778: Status 404 returned error can't find the container with id 772bed82dcb1aa07361825b7f6893d507fcfa74c070d6f5df70ec5be90c33778 Dec 06 05:59:01 crc kubenswrapper[4957]: I1206 05:59:01.014397 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" event={"ID":"e956383f-8ae1-444a-b876-a85b2bcd1bcb","Type":"ContainerStarted","Data":"772bed82dcb1aa07361825b7f6893d507fcfa74c070d6f5df70ec5be90c33778"} Dec 06 05:59:01 crc kubenswrapper[4957]: I1206 05:59:01.261758 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h"] Dec 06 05:59:01 crc kubenswrapper[4957]: W1206 05:59:01.281020 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb48eed1d_ba6d_41a4_bb23_2d6cc39c4e2e.slice/crio-36e5db46a062e44d5c017bbfac74563b71114b16cfcfe518ccc3acb1ac8fcd45 WatchSource:0}: Error finding container 36e5db46a062e44d5c017bbfac74563b71114b16cfcfe518ccc3acb1ac8fcd45: Status 404 returned error can't find the container with id 36e5db46a062e44d5c017bbfac74563b71114b16cfcfe518ccc3acb1ac8fcd45 Dec 06 05:59:01 crc kubenswrapper[4957]: I1206 05:59:01.300276 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8"] Dec 06 05:59:01 crc kubenswrapper[4957]: W1206 05:59:01.314981 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9a79638_2a84_4bc8_a3c8_3289f3a1adae.slice/crio-88bc33123c2041e5ce0478d48fa7b5281cc73e76af8bf69eade30f7eda97fa51 WatchSource:0}: Error finding container 88bc33123c2041e5ce0478d48fa7b5281cc73e76af8bf69eade30f7eda97fa51: Status 404 returned error can't find the container with id 88bc33123c2041e5ce0478d48fa7b5281cc73e76af8bf69eade30f7eda97fa51 Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.023061 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" event={"ID":"6276ae06-7d72-409a-81f2-70ba70bf3d86","Type":"ContainerStarted","Data":"8fc3840b45d58ba42348df1cc8e4145064d05ad546e6760d09b419188868ae11"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.024448 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" event={"ID":"bf2466dc-9cd3-4be7-9d17-846eccda13a8","Type":"ContainerStarted","Data":"5076d30244d960d404b6422004dfa404b5ee48e33c55467ae7f9b9ec8cf580bd"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.026751 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" event={"ID":"4affc2fc-c552-4a1b-b59e-31c7ff2cbb87","Type":"ContainerStarted","Data":"bb2aebf948ccc652202fbb5b8117769701ac377ae307b38a98efe03c6406552a"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.028289 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" event={"ID":"ea50a02a-23dc-4fcf-bcfd-048b66ce8152","Type":"ContainerStarted","Data":"e19d9ddac8d2f90fbb68b525017298cb045909d30e6f04ba275b402b14261822"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.029388 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" event={"ID":"e7ea0507-7466-4c88-a20b-2c936c6cdccb","Type":"ContainerStarted","Data":"c7123a7274414fd79880f1b497a02a499a95db450bcc973daa8ef25fcb91c58b"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.030921 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" event={"ID":"b87d0440-b095-4540-a0a4-cb71461be4bd","Type":"ContainerStarted","Data":"7c0445c71446193df64d9c692bea05b08d8211467babf1fda969e16c44eb4f42"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.032816 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" event={"ID":"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e","Type":"ContainerStarted","Data":"36e5db46a062e44d5c017bbfac74563b71114b16cfcfe518ccc3acb1ac8fcd45"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.034032 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" event={"ID":"7aa1c48d-537b-4674-a284-1d929fc9cd16","Type":"ContainerStarted","Data":"4d445aee549b00e279e84d9199962fbadc35cf3cf2511fb11fad56c42f7b8a68"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.059961 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" event={"ID":"db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5","Type":"ContainerStarted","Data":"41d70f7abae2b16f31ddfdb9f3c3a5856b4d9513e86ce7682ad5c8e2f9427d7f"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.064324 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" event={"ID":"f9a79638-2a84-4bc8-a3c8-3289f3a1adae","Type":"ContainerStarted","Data":"88bc33123c2041e5ce0478d48fa7b5281cc73e76af8bf69eade30f7eda97fa51"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.068005 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" event={"ID":"4de58666-a4e1-4cae-99d7-2088192c51db","Type":"ContainerStarted","Data":"e5ef3fb43ff1b4f5a5a9c0b75ecefba9e55fab60dea0b5ac24a91b5e0e4717c9"} Dec 06 05:59:02 crc kubenswrapper[4957]: I1206 05:59:02.070594 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" event={"ID":"ee5ec7b9-749f-4b24-826c-bebb580f7c5b","Type":"ContainerStarted","Data":"98488a56786c63748ea2050151613e1a22c24862599caaabaf7d87d7016c3546"} Dec 06 05:59:06 crc kubenswrapper[4957]: I1206 05:59:06.131265 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" event={"ID":"b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e","Type":"ContainerStarted","Data":"1ae3407ce4bcc821e83599fc16f7b71559871656921bc8dba02822be88bcc0dc"} Dec 06 05:59:06 crc kubenswrapper[4957]: I1206 05:59:06.131820 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:59:06 crc kubenswrapper[4957]: I1206 05:59:06.133782 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" event={"ID":"d349e041-2168-4b8a-b6c4-a5840de8c9d8","Type":"ContainerStarted","Data":"e1b4e79b699ea4134f2e149f96a9567ae92530ea46f9824199ee01e3cc29d193"} Dec 06 05:59:06 crc kubenswrapper[4957]: I1206 05:59:06.137343 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" event={"ID":"47c9992f-497a-4d57-8c37-c344ad49c199","Type":"ContainerStarted","Data":"88720cfee5968a79e45547f3b5d01b1b33134abd111e8a74f1636d36b4312e74"} Dec 06 05:59:06 crc kubenswrapper[4957]: I1206 05:59:06.139823 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" event={"ID":"0a9ebb33-9805-4dec-a879-df7417c8ce7b","Type":"ContainerStarted","Data":"39a7ada4b7c63827ef83404d05ba96637d8ada76c4cae82820b0e002f44eb52d"} Dec 06 05:59:06 crc kubenswrapper[4957]: I1206 05:59:06.167532 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" podStartSLOduration=41.167511264 podStartE2EDuration="41.167511264s" podCreationTimestamp="2025-12-06 05:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 05:59:06.160554673 +0000 UTC m=+1185.810822325" watchObservedRunningTime="2025-12-06 05:59:06.167511264 +0000 UTC m=+1185.817778896" Dec 06 05:59:10 crc kubenswrapper[4957]: I1206 05:59:10.180579 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:59:10 crc kubenswrapper[4957]: I1206 05:59:10.181058 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:59:13 crc kubenswrapper[4957]: I1206 05:59:13.070246 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-75c7c544f9-bnr4h" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.781523 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.782308 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bq4nj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-d7r5m_openstack-operators(c5d6f879-3842-4481-88bd-bb090eecd9fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.783568 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" podUID="c5d6f879-3842-4481-88bd-bb090eecd9fc" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.850470 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.850678 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2dg42,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-rs7b8_openstack-operators(ce46ec44-fc83-46e0-9f85-159e05452f0f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.852314 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" podUID="ce46ec44-fc83-46e0-9f85-159e05452f0f" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.860894 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.861084 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kd86h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-rrq65_openstack-operators(33674637-d0b9-40c9-a482-f4f18c39a80d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.863346 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" podUID="33674637-d0b9-40c9-a482-f4f18c39a80d" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.960126 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.960467 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lxk6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-xjbq2_openstack-operators(2d7a7ff4-095a-4dff-9c96-25d03888a51e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.961666 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" podUID="2d7a7ff4-095a-4dff-9c96-25d03888a51e" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.964226 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.964389 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6kt6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-88lgf_openstack-operators(5904f94f-274a-4b7f-92be-5b8acc5736ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:33 crc kubenswrapper[4957]: E1206 05:59:33.965779 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" podUID="5904f94f-274a-4b7f-92be-5b8acc5736ab" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.063885 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.064055 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q84xh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-nr7jl_openstack-operators(f2ab1e95-3e28-4d99-9340-f1eb24bc6dca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.065291 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" podUID="f2ab1e95-3e28-4d99-9340-f1eb24bc6dca" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.381783 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.381788 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.381957 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j4zc8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-mzk64_openstack-operators(bf2466dc-9cd3-4be7-9d17-846eccda13a8): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.382068 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cq2m2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-fvj4l_openstack-operators(ea50a02a-23dc-4fcf-bcfd-048b66ce8152): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.383598 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" podUID="bf2466dc-9cd3-4be7-9d17-846eccda13a8" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.383717 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" podUID="ea50a02a-23dc-4fcf-bcfd-048b66ce8152" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.384067 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.384183 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5kxj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-w7wgp_openstack-operators(7aa1c48d-537b-4674-a284-1d929fc9cd16): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.384326 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.384603 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z4h5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-9m6rj_openstack-operators(e7ea0507-7466-4c88-a20b-2c936c6cdccb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.384869 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.385570 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" podUID="7aa1c48d-537b-4674-a284-1d929fc9cd16" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.385328 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5fnxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8_openstack-operators(f9a79638-2a84-4bc8-a3c8-3289f3a1adae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.385708 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" podUID="e7ea0507-7466-4c88-a20b-2c936c6cdccb" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.387407 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.387491 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.387626 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twmvf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-k4mm8_openstack-operators(ee5ec7b9-749f-4b24-826c-bebb580f7c5b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.387648 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5bzlc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-2cx79_openstack-operators(4de58666-a4e1-4cae-99d7-2088192c51db): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.389390 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.389390 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" podUID="4de58666-a4e1-4cae-99d7-2088192c51db" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.389517 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4vcgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-7nwnc_openstack-operators(4affc2fc-c552-4a1b-b59e-31c7ff2cbb87): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.389529 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" podUID="ee5ec7b9-749f-4b24-826c-bebb580f7c5b" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.391423 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" podUID="4affc2fc-c552-4a1b-b59e-31c7ff2cbb87" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.393553 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.393686 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tc2ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-c2t2c_openstack-operators(db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.394927 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" podUID="db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.395329 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.395419 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gfdsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-rpf52_openstack-operators(6276ae06-7d72-409a-81f2-70ba70bf3d86): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.396607 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" podUID="6276ae06-7d72-409a-81f2-70ba70bf3d86" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.408120 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.408312 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t8lz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-8mtps_openstack-operators(b87d0440-b095-4540-a0a4-cb71461be4bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 05:59:34 crc kubenswrapper[4957]: E1206 05:59:34.409496 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" podUID="b87d0440-b095-4540-a0a4-cb71461be4bd" Dec 06 05:59:35 crc kubenswrapper[4957]: E1206 05:59:35.366434 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" podUID="f9a79638-2a84-4bc8-a3c8-3289f3a1adae" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.409352 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" event={"ID":"31332329-9dd4-4665-b1d0-9662ee3d1b8e","Type":"ContainerStarted","Data":"a3fcfb3300cba6cb5639e130785edce492d27177997cf1244dee5a625e9571c2"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.416870 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" event={"ID":"33674637-d0b9-40c9-a482-f4f18c39a80d","Type":"ContainerStarted","Data":"f1eced4aa8045b983aa493af17b030bdc6984ef2b65b8ecca66c2ef4837bc49d"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.430170 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" event={"ID":"0a9ebb33-9805-4dec-a879-df7417c8ce7b","Type":"ContainerStarted","Data":"b6d40269fd58124a058485bad52621811335291a76f61e501d994f38013b7305"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.431159 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.433540 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.448376 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6gccn" podStartSLOduration=38.650048657 podStartE2EDuration="1m10.448358233s" podCreationTimestamp="2025-12-06 05:58:25 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.51122838 +0000 UTC m=+1149.161496012" lastFinishedPulling="2025-12-06 05:59:01.309537956 +0000 UTC m=+1180.959805588" observedRunningTime="2025-12-06 05:59:35.438642759 +0000 UTC m=+1215.088910391" watchObservedRunningTime="2025-12-06 05:59:35.448358233 +0000 UTC m=+1215.098625865" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.452405 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" event={"ID":"f9a79638-2a84-4bc8-a3c8-3289f3a1adae","Type":"ContainerStarted","Data":"fb7c292eebbffddc0acc0f298f5c5b53406fbe149782168cd2e409bc3213d60f"} Dec 06 05:59:35 crc kubenswrapper[4957]: E1206 05:59:35.462460 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" podUID="f9a79638-2a84-4bc8-a3c8-3289f3a1adae" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.469147 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" event={"ID":"47c9992f-497a-4d57-8c37-c344ad49c199","Type":"ContainerStarted","Data":"bd6b1bfe10cc1615e3d4dd494972ba33aefbb0b164bf6693876395c2731ebf10"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.469694 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mc7wf" podStartSLOduration=6.413701799 podStartE2EDuration="1m11.469673286s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.478378772 +0000 UTC m=+1149.128646404" lastFinishedPulling="2025-12-06 05:59:34.534350259 +0000 UTC m=+1214.184617891" observedRunningTime="2025-12-06 05:59:35.461570035 +0000 UTC m=+1215.111837687" watchObservedRunningTime="2025-12-06 05:59:35.469673286 +0000 UTC m=+1215.119940918" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.469748 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.475657 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.480026 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" event={"ID":"2d7a7ff4-095a-4dff-9c96-25d03888a51e","Type":"ContainerStarted","Data":"6d2a320a9610cdb18af31c90d2770d1ffb0304bf506358c7005db8e399ccc3ba"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.492616 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" event={"ID":"5904f94f-274a-4b7f-92be-5b8acc5736ab","Type":"ContainerStarted","Data":"db88e4b355d18a4e1a0b81cdaa416092b4e2b780eaa1c78b667ebb671caff4c2"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.528729 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" event={"ID":"c5d6f879-3842-4481-88bd-bb090eecd9fc","Type":"ContainerStarted","Data":"a2f4846797a20eb818f9bff95f1a4dc5d00b4c999a63eb9be08392ca1110fb46"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.529600 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.539184 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2x6nl" podStartSLOduration=6.497489103 podStartE2EDuration="1m11.539171113s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.494711404 +0000 UTC m=+1149.144979036" lastFinishedPulling="2025-12-06 05:59:34.536393414 +0000 UTC m=+1214.186661046" observedRunningTime="2025-12-06 05:59:35.528552434 +0000 UTC m=+1215.178820076" watchObservedRunningTime="2025-12-06 05:59:35.539171113 +0000 UTC m=+1215.189438745" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.544448 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" event={"ID":"f2ab1e95-3e28-4d99-9340-f1eb24bc6dca","Type":"ContainerStarted","Data":"a179d90a981e4ab3ab9205a800a01536fefc5ad04a207a7a9d9378a2b392d66e"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.586489 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" event={"ID":"d349e041-2168-4b8a-b6c4-a5840de8c9d8","Type":"ContainerStarted","Data":"e4f2bb90b58f6d038c66c95b9e06576bbaa255cfca2e4226d24c61c61ee32978"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.587074 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.596562 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.615419 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" event={"ID":"e956383f-8ae1-444a-b876-a85b2bcd1bcb","Type":"ContainerStarted","Data":"1348b8128039c68d5a0799a3320c322e34fb6054e52de7786fc2de5d10924f50"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.615464 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" event={"ID":"e956383f-8ae1-444a-b876-a85b2bcd1bcb","Type":"ContainerStarted","Data":"d673af04a1109def5542a42a907740cc7da7daa3f80e7f408afad4c41adaa533"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.616117 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.625161 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" event={"ID":"ce46ec44-fc83-46e0-9f85-159e05452f0f","Type":"ContainerStarted","Data":"c051e957d5fee0e2c22d30c9bd22ee6130fb61a7aea5389616a47586e44c2da9"} Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628137 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628167 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628179 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628190 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628200 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628211 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628415 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628638 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628674 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.628928 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.644180 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.645075 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-g6zcq" podStartSLOduration=6.611119046 podStartE2EDuration="1m11.645059776s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.498541207 +0000 UTC m=+1149.148808839" lastFinishedPulling="2025-12-06 05:59:34.532481937 +0000 UTC m=+1214.182749569" observedRunningTime="2025-12-06 05:59:35.643433052 +0000 UTC m=+1215.293700684" watchObservedRunningTime="2025-12-06 05:59:35.645059776 +0000 UTC m=+1215.295327408" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.645712 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.646080 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.646158 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.646213 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.646253 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.646421 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.646479 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.647357 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.648094 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.650093 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" podStartSLOduration=5.684449578 podStartE2EDuration="1m11.650076173s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.002456701 +0000 UTC m=+1148.652724333" lastFinishedPulling="2025-12-06 05:59:34.968083296 +0000 UTC m=+1214.618350928" observedRunningTime="2025-12-06 05:59:35.573024488 +0000 UTC m=+1215.223292130" watchObservedRunningTime="2025-12-06 05:59:35.650076173 +0000 UTC m=+1215.300343805" Dec 06 05:59:35 crc kubenswrapper[4957]: I1206 05:59:35.820166 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" podStartSLOduration=41.15576383 podStartE2EDuration="1m11.820147899s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:59:00.452679342 +0000 UTC m=+1180.102946974" lastFinishedPulling="2025-12-06 05:59:31.117063381 +0000 UTC m=+1210.767331043" observedRunningTime="2025-12-06 05:59:35.816386285 +0000 UTC m=+1215.466653917" watchObservedRunningTime="2025-12-06 05:59:35.820147899 +0000 UTC m=+1215.470415531" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.642002 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" event={"ID":"ea50a02a-23dc-4fcf-bcfd-048b66ce8152","Type":"ContainerStarted","Data":"95dec5c0dcf9ced9d4ea3959cf6229a87eefba2d5bb1032d4edb7277f5759c4f"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.654018 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" event={"ID":"2d7a7ff4-095a-4dff-9c96-25d03888a51e","Type":"ContainerStarted","Data":"402be7e7e33d0b00934efa59d89f2029e129777b90d00e3cc4ae172c8295b0b3"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.654343 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.676589 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" event={"ID":"5904f94f-274a-4b7f-92be-5b8acc5736ab","Type":"ContainerStarted","Data":"c7769886bf67e98450289399c2aaa7e6533394775d58b6a01230d71f0e278147"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.676663 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" event={"ID":"f2ab1e95-3e28-4d99-9340-f1eb24bc6dca","Type":"ContainerStarted","Data":"fc81f6a3a50ccca9ce7c0a7ca25fcad0da0fbed7a226d1d02375cfe75b7e8441"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.676679 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" event={"ID":"db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5","Type":"ContainerStarted","Data":"cb3b704c7e0093fc55126ea0d995d6f61c89a25a198ef87fb553b402e13f506c"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.676691 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" event={"ID":"33674637-d0b9-40c9-a482-f4f18c39a80d","Type":"ContainerStarted","Data":"d47fc31b26a1df2ac2d41b79769125a5fa05711fe9416db87e633968792e2a71"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.676731 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.676746 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.676758 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.689084 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" event={"ID":"4de58666-a4e1-4cae-99d7-2088192c51db","Type":"ContainerStarted","Data":"e27373134502428db391c6d14d8ae0b413aad2d8c9c8aa2bc47be874f5e83d64"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.690162 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fvj4l" podStartSLOduration=50.14360733 podStartE2EDuration="1m12.690148681s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.454550458 +0000 UTC m=+1149.104818090" lastFinishedPulling="2025-12-06 05:58:52.001091809 +0000 UTC m=+1171.651359441" observedRunningTime="2025-12-06 05:59:36.687898969 +0000 UTC m=+1216.338166631" watchObservedRunningTime="2025-12-06 05:59:36.690148681 +0000 UTC m=+1216.340416313" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.705250 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" event={"ID":"bf2466dc-9cd3-4be7-9d17-846eccda13a8","Type":"ContainerStarted","Data":"26311e5d476944609ab866dc299feb8eb2d30c89e373236e618a84ce9d9728ab"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.722649 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" event={"ID":"4affc2fc-c552-4a1b-b59e-31c7ff2cbb87","Type":"ContainerStarted","Data":"9b373344e793cd42663d78dedaaaca6387f251a9b939d3187a124302e57e6c85"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.734173 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" event={"ID":"ee5ec7b9-749f-4b24-826c-bebb580f7c5b","Type":"ContainerStarted","Data":"bf3fc0f34ec859249a78ef4f2e9e2c1a81da979e9459c3eeb07970b40d3072fd"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.755292 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" event={"ID":"ce46ec44-fc83-46e0-9f85-159e05452f0f","Type":"ContainerStarted","Data":"35a74f8255a1a17b19392b22180368e6b3e318f262924a7018762f0c7713cfa9"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.756109 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" podStartSLOduration=7.04574137 podStartE2EDuration="1m12.756093052s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.452323438 +0000 UTC m=+1149.102591060" lastFinishedPulling="2025-12-06 05:59:35.16267511 +0000 UTC m=+1214.812942742" observedRunningTime="2025-12-06 05:59:36.754783336 +0000 UTC m=+1216.405050988" watchObservedRunningTime="2025-12-06 05:59:36.756093052 +0000 UTC m=+1216.406360694" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.756164 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.757557 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" podStartSLOduration=7.220032588 podStartE2EDuration="1m12.757548972s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.458937497 +0000 UTC m=+1149.109205119" lastFinishedPulling="2025-12-06 05:59:34.996453871 +0000 UTC m=+1214.646721503" observedRunningTime="2025-12-06 05:59:36.712220484 +0000 UTC m=+1216.362488136" watchObservedRunningTime="2025-12-06 05:59:36.757548972 +0000 UTC m=+1216.407816614" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.800164 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" event={"ID":"c5d6f879-3842-4481-88bd-bb090eecd9fc","Type":"ContainerStarted","Data":"d744ff8efcfe3fd04a8cc07e2b7eba9d5f6f5e52001551ed7c7a43dda376c663"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.824700 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" event={"ID":"6276ae06-7d72-409a-81f2-70ba70bf3d86","Type":"ContainerStarted","Data":"edd0ded2e5acb19cf6a399f938cb06a25cfbc3df0c9fc69bb681fa931ad2674b"} Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.826553 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" podStartSLOduration=6.845982559 podStartE2EDuration="1m12.826536326s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.01649817 +0000 UTC m=+1148.666765802" lastFinishedPulling="2025-12-06 05:59:34.997051937 +0000 UTC m=+1214.647319569" observedRunningTime="2025-12-06 05:59:36.79591657 +0000 UTC m=+1216.446184222" watchObservedRunningTime="2025-12-06 05:59:36.826536326 +0000 UTC m=+1216.476803958" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.829627 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" podStartSLOduration=7.298424068 podStartE2EDuration="1m12.82962091s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.464168228 +0000 UTC m=+1149.114435860" lastFinishedPulling="2025-12-06 05:59:34.99536507 +0000 UTC m=+1214.645632702" observedRunningTime="2025-12-06 05:59:36.818859456 +0000 UTC m=+1216.469127098" watchObservedRunningTime="2025-12-06 05:59:36.82962091 +0000 UTC m=+1216.479888542" Dec 06 05:59:36 crc kubenswrapper[4957]: E1206 05:59:36.832206 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" podUID="f9a79638-2a84-4bc8-a3c8-3289f3a1adae" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.852404 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c2t2c" podStartSLOduration=50.242369541 podStartE2EDuration="1m12.852379542s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.406971812 +0000 UTC m=+1149.057239444" lastFinishedPulling="2025-12-06 05:58:52.016981813 +0000 UTC m=+1171.667249445" observedRunningTime="2025-12-06 05:59:36.851597091 +0000 UTC m=+1216.501864733" watchObservedRunningTime="2025-12-06 05:59:36.852379542 +0000 UTC m=+1216.502647184" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.889841 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k4mm8" podStartSLOduration=44.883798076 podStartE2EDuration="1m11.889805895s" podCreationTimestamp="2025-12-06 05:58:25 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.475909966 +0000 UTC m=+1149.126177598" lastFinishedPulling="2025-12-06 05:58:56.481917795 +0000 UTC m=+1176.132185417" observedRunningTime="2025-12-06 05:59:36.881708704 +0000 UTC m=+1216.531976336" watchObservedRunningTime="2025-12-06 05:59:36.889805895 +0000 UTC m=+1216.540073527" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.914112 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rpf52" podStartSLOduration=50.366312048 podStartE2EDuration="1m12.914095108s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.472859903 +0000 UTC m=+1149.123127535" lastFinishedPulling="2025-12-06 05:58:52.020642963 +0000 UTC m=+1171.670910595" observedRunningTime="2025-12-06 05:59:36.91124421 +0000 UTC m=+1216.561511842" watchObservedRunningTime="2025-12-06 05:59:36.914095108 +0000 UTC m=+1216.564362740" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.967543 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mzk64" podStartSLOduration=50.359439058 podStartE2EDuration="1m12.967519967s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.413933301 +0000 UTC m=+1149.064200933" lastFinishedPulling="2025-12-06 05:58:52.02201421 +0000 UTC m=+1171.672281842" observedRunningTime="2025-12-06 05:59:36.942130653 +0000 UTC m=+1216.592398305" watchObservedRunningTime="2025-12-06 05:59:36.967519967 +0000 UTC m=+1216.617787599" Dec 06 05:59:36 crc kubenswrapper[4957]: I1206 05:59:36.993504 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" podStartSLOduration=7.379130597 podStartE2EDuration="1m12.993484537s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.382847071 +0000 UTC m=+1149.033114703" lastFinishedPulling="2025-12-06 05:59:34.997200971 +0000 UTC m=+1214.647468643" observedRunningTime="2025-12-06 05:59:36.989415455 +0000 UTC m=+1216.639683097" watchObservedRunningTime="2025-12-06 05:59:36.993484537 +0000 UTC m=+1216.643752169" Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.030942 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7nwnc" podStartSLOduration=41.951476744 podStartE2EDuration="1m13.030921659s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.494664073 +0000 UTC m=+1149.144931705" lastFinishedPulling="2025-12-06 05:59:00.574108988 +0000 UTC m=+1180.224376620" observedRunningTime="2025-12-06 05:59:37.02473145 +0000 UTC m=+1216.674999102" watchObservedRunningTime="2025-12-06 05:59:37.030921659 +0000 UTC m=+1216.681189301" Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.060176 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2cx79" podStartSLOduration=46.009901961 podStartE2EDuration="1m13.060156107s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.431159906 +0000 UTC m=+1149.081427548" lastFinishedPulling="2025-12-06 05:58:56.481414062 +0000 UTC m=+1176.131681694" observedRunningTime="2025-12-06 05:59:37.053238768 +0000 UTC m=+1216.703506420" watchObservedRunningTime="2025-12-06 05:59:37.060156107 +0000 UTC m=+1216.710423749" Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.840303 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" event={"ID":"e7ea0507-7466-4c88-a20b-2c936c6cdccb","Type":"ContainerStarted","Data":"fd72af162cb0b106774990d7e434d4dceb2a64afc9c21c7a4ec8bb971c525daf"} Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.842559 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" event={"ID":"b87d0440-b095-4540-a0a4-cb71461be4bd","Type":"ContainerStarted","Data":"1a845c7eecd3e36fa47727f24f1125b0ccbb9bf5e0c490d2f87d2d9de14e6593"} Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.845615 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" event={"ID":"7aa1c48d-537b-4674-a284-1d929fc9cd16","Type":"ContainerStarted","Data":"aa426958da5414fe5c0d0fbaa404558e2ac4757d09b70a99d996ecbae88919a4"} Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.867960 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-9m6rj" podStartSLOduration=50.482495783 podStartE2EDuration="1m13.867935551s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:28.635190895 +0000 UTC m=+1148.285458537" lastFinishedPulling="2025-12-06 05:58:52.020630673 +0000 UTC m=+1171.670898305" observedRunningTime="2025-12-06 05:59:37.858205944 +0000 UTC m=+1217.508473596" watchObservedRunningTime="2025-12-06 05:59:37.867935551 +0000 UTC m=+1217.518203183" Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.890575 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7wgp" podStartSLOduration=42.821833178 podStartE2EDuration="1m13.890555728s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.482243966 +0000 UTC m=+1149.132511598" lastFinishedPulling="2025-12-06 05:59:00.550966506 +0000 UTC m=+1180.201234148" observedRunningTime="2025-12-06 05:59:37.887010791 +0000 UTC m=+1217.537278443" watchObservedRunningTime="2025-12-06 05:59:37.890555728 +0000 UTC m=+1217.540823380" Dec 06 05:59:37 crc kubenswrapper[4957]: I1206 05:59:37.920236 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-8mtps" podStartSLOduration=42.826445515 podStartE2EDuration="1m13.920218088s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:58:29.478949508 +0000 UTC m=+1149.129217140" lastFinishedPulling="2025-12-06 05:59:00.572722061 +0000 UTC m=+1180.222989713" observedRunningTime="2025-12-06 05:59:37.918970184 +0000 UTC m=+1217.569237836" watchObservedRunningTime="2025-12-06 05:59:37.920218088 +0000 UTC m=+1217.570485720" Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.180737 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.181231 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.181303 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.182365 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3f75ddbe4ecf31b2a5519fb70a8ed4815614284758f53e65c5fc0533e95c0a3"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.182502 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://f3f75ddbe4ecf31b2a5519fb70a8ed4815614284758f53e65c5fc0533e95c0a3" gracePeriod=600 Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.880747 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="f3f75ddbe4ecf31b2a5519fb70a8ed4815614284758f53e65c5fc0533e95c0a3" exitCode=0 Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.881260 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"f3f75ddbe4ecf31b2a5519fb70a8ed4815614284758f53e65c5fc0533e95c0a3"} Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.881572 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"b3b7a79d60429782e416177d4e1df59e205a0e6bbe183ad64871d22be42eb374"} Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.881598 4957 scope.go:117] "RemoveContainer" containerID="c315dd48c0a384ca90d94b9ce0d09121f3d64da936f2696223ce8145d1b5a516" Dec 06 05:59:40 crc kubenswrapper[4957]: I1206 05:59:40.956160 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9bc586d8b-5f5p9" Dec 06 05:59:44 crc kubenswrapper[4957]: I1206 05:59:44.926787 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d7r5m" Dec 06 05:59:44 crc kubenswrapper[4957]: I1206 05:59:44.955111 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-rs7b8" Dec 06 05:59:45 crc kubenswrapper[4957]: I1206 05:59:45.036625 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-88lgf" Dec 06 05:59:45 crc kubenswrapper[4957]: I1206 05:59:45.640407 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-rrq65" Dec 06 05:59:45 crc kubenswrapper[4957]: I1206 05:59:45.833693 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-xjbq2" Dec 06 05:59:46 crc kubenswrapper[4957]: I1206 05:59:46.153241 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-nr7jl" Dec 06 05:59:49 crc kubenswrapper[4957]: I1206 05:59:49.954813 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" event={"ID":"f9a79638-2a84-4bc8-a3c8-3289f3a1adae","Type":"ContainerStarted","Data":"9b2b64a3eab52607e6022de32068ddebda458005102bf74a18b430f1d7222c56"} Dec 06 05:59:49 crc kubenswrapper[4957]: I1206 05:59:49.956526 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 05:59:49 crc kubenswrapper[4957]: I1206 05:59:49.987500 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" podStartSLOduration=38.145891779 podStartE2EDuration="1m25.987480347s" podCreationTimestamp="2025-12-06 05:58:24 +0000 UTC" firstStartedPulling="2025-12-06 05:59:01.320166305 +0000 UTC m=+1180.970433937" lastFinishedPulling="2025-12-06 05:59:49.161754873 +0000 UTC m=+1228.812022505" observedRunningTime="2025-12-06 05:59:49.984332381 +0000 UTC m=+1229.634600033" watchObservedRunningTime="2025-12-06 05:59:49.987480347 +0000 UTC m=+1229.637747979" Dec 06 05:59:58 crc kubenswrapper[4957]: I1206 05:59:58.928020 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.158498 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk"] Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.160246 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.176125 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.229026 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.230381 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk"] Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.254958 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4998358d-ec92-4690-b968-23c82a67a764-config-volume\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.255200 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4998358d-ec92-4690-b968-23c82a67a764-secret-volume\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.255350 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbn5v\" (UniqueName: \"kubernetes.io/projected/4998358d-ec92-4690-b968-23c82a67a764-kube-api-access-tbn5v\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.358012 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4998358d-ec92-4690-b968-23c82a67a764-secret-volume\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.358120 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbn5v\" (UniqueName: \"kubernetes.io/projected/4998358d-ec92-4690-b968-23c82a67a764-kube-api-access-tbn5v\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.358215 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4998358d-ec92-4690-b968-23c82a67a764-config-volume\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.360261 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4998358d-ec92-4690-b968-23c82a67a764-config-volume\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.369973 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4998358d-ec92-4690-b968-23c82a67a764-secret-volume\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.382216 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbn5v\" (UniqueName: \"kubernetes.io/projected/4998358d-ec92-4690-b968-23c82a67a764-kube-api-access-tbn5v\") pod \"collect-profiles-29416680-tn5dk\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:00 crc kubenswrapper[4957]: I1206 06:00:00.541206 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:01 crc kubenswrapper[4957]: I1206 06:00:01.029146 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk"] Dec 06 06:00:01 crc kubenswrapper[4957]: I1206 06:00:01.055291 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" event={"ID":"4998358d-ec92-4690-b968-23c82a67a764","Type":"ContainerStarted","Data":"0e6a3f6948dca54b640482800f495993fc04bc9ede5742a582bc539f81ffac42"} Dec 06 06:00:02 crc kubenswrapper[4957]: I1206 06:00:02.069211 4957 generic.go:334] "Generic (PLEG): container finished" podID="4998358d-ec92-4690-b968-23c82a67a764" containerID="e9fc2079c75ee4eeed9992d206edac0242fdb11945be37a78d6b98e4e4884ebd" exitCode=0 Dec 06 06:00:02 crc kubenswrapper[4957]: I1206 06:00:02.069286 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" event={"ID":"4998358d-ec92-4690-b968-23c82a67a764","Type":"ContainerDied","Data":"e9fc2079c75ee4eeed9992d206edac0242fdb11945be37a78d6b98e4e4884ebd"} Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.413885 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.509818 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4998358d-ec92-4690-b968-23c82a67a764-config-volume\") pod \"4998358d-ec92-4690-b968-23c82a67a764\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.509886 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbn5v\" (UniqueName: \"kubernetes.io/projected/4998358d-ec92-4690-b968-23c82a67a764-kube-api-access-tbn5v\") pod \"4998358d-ec92-4690-b968-23c82a67a764\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.509956 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4998358d-ec92-4690-b968-23c82a67a764-secret-volume\") pod \"4998358d-ec92-4690-b968-23c82a67a764\" (UID: \"4998358d-ec92-4690-b968-23c82a67a764\") " Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.510618 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4998358d-ec92-4690-b968-23c82a67a764-config-volume" (OuterVolumeSpecName: "config-volume") pod "4998358d-ec92-4690-b968-23c82a67a764" (UID: "4998358d-ec92-4690-b968-23c82a67a764"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.514851 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4998358d-ec92-4690-b968-23c82a67a764-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4998358d-ec92-4690-b968-23c82a67a764" (UID: "4998358d-ec92-4690-b968-23c82a67a764"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.515197 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4998358d-ec92-4690-b968-23c82a67a764-kube-api-access-tbn5v" (OuterVolumeSpecName: "kube-api-access-tbn5v") pod "4998358d-ec92-4690-b968-23c82a67a764" (UID: "4998358d-ec92-4690-b968-23c82a67a764"). InnerVolumeSpecName "kube-api-access-tbn5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.611800 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4998358d-ec92-4690-b968-23c82a67a764-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.611865 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbn5v\" (UniqueName: \"kubernetes.io/projected/4998358d-ec92-4690-b968-23c82a67a764-kube-api-access-tbn5v\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:03 crc kubenswrapper[4957]: I1206 06:00:03.611881 4957 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4998358d-ec92-4690-b968-23c82a67a764-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:04 crc kubenswrapper[4957]: I1206 06:00:04.098702 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" event={"ID":"4998358d-ec92-4690-b968-23c82a67a764","Type":"ContainerDied","Data":"0e6a3f6948dca54b640482800f495993fc04bc9ede5742a582bc539f81ffac42"} Dec 06 06:00:04 crc kubenswrapper[4957]: I1206 06:00:04.098736 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e6a3f6948dca54b640482800f495993fc04bc9ede5742a582bc539f81ffac42" Dec 06 06:00:04 crc kubenswrapper[4957]: I1206 06:00:04.098793 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.477336 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tlnjq"] Dec 06 06:00:13 crc kubenswrapper[4957]: E1206 06:00:13.479891 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4998358d-ec92-4690-b968-23c82a67a764" containerName="collect-profiles" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.480022 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="4998358d-ec92-4690-b968-23c82a67a764" containerName="collect-profiles" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.480434 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="4998358d-ec92-4690-b968-23c82a67a764" containerName="collect-profiles" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.483030 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.487207 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.487331 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.487954 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-55h9n" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.488017 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.488145 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tlnjq"] Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.573765 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gdqct"] Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.587303 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.593322 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.611261 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gdqct"] Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.682975 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.683065 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c48c03-e899-475e-aaad-5db38c19c26e-config\") pod \"dnsmasq-dns-675f4bcbfc-tlnjq\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.683110 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-config\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.683133 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dc7j\" (UniqueName: \"kubernetes.io/projected/e8c48c03-e899-475e-aaad-5db38c19c26e-kube-api-access-8dc7j\") pod \"dnsmasq-dns-675f4bcbfc-tlnjq\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.683160 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bml9s\" (UniqueName: \"kubernetes.io/projected/a61521c5-8048-4bf1-be5c-6d84f0a915c0-kube-api-access-bml9s\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.784378 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.784481 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c48c03-e899-475e-aaad-5db38c19c26e-config\") pod \"dnsmasq-dns-675f4bcbfc-tlnjq\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.784518 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-config\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.784538 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dc7j\" (UniqueName: \"kubernetes.io/projected/e8c48c03-e899-475e-aaad-5db38c19c26e-kube-api-access-8dc7j\") pod \"dnsmasq-dns-675f4bcbfc-tlnjq\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.784571 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bml9s\" (UniqueName: \"kubernetes.io/projected/a61521c5-8048-4bf1-be5c-6d84f0a915c0-kube-api-access-bml9s\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.785748 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.785812 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-config\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.785871 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c48c03-e899-475e-aaad-5db38c19c26e-config\") pod \"dnsmasq-dns-675f4bcbfc-tlnjq\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.807307 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bml9s\" (UniqueName: \"kubernetes.io/projected/a61521c5-8048-4bf1-be5c-6d84f0a915c0-kube-api-access-bml9s\") pod \"dnsmasq-dns-78dd6ddcc-gdqct\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.807713 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dc7j\" (UniqueName: \"kubernetes.io/projected/e8c48c03-e899-475e-aaad-5db38c19c26e-kube-api-access-8dc7j\") pod \"dnsmasq-dns-675f4bcbfc-tlnjq\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:13 crc kubenswrapper[4957]: I1206 06:00:13.931279 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:14 crc kubenswrapper[4957]: I1206 06:00:14.106977 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:14 crc kubenswrapper[4957]: I1206 06:00:14.349308 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gdqct"] Dec 06 06:00:14 crc kubenswrapper[4957]: I1206 06:00:14.523282 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tlnjq"] Dec 06 06:00:15 crc kubenswrapper[4957]: I1206 06:00:15.203183 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" event={"ID":"e8c48c03-e899-475e-aaad-5db38c19c26e","Type":"ContainerStarted","Data":"17cf687f6725ad87ba5a869a2011d2eb41340e7535d658e735e42093e1d5218e"} Dec 06 06:00:15 crc kubenswrapper[4957]: I1206 06:00:15.204285 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" event={"ID":"a61521c5-8048-4bf1-be5c-6d84f0a915c0","Type":"ContainerStarted","Data":"2fe24e8ae164a735ab666614e944efcc5c1832e28e4f8737db55a61a00c214a8"} Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.603361 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tlnjq"] Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.624896 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jgxdf"] Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.626116 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.653869 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jgxdf"] Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.727696 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.727761 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn9tw\" (UniqueName: \"kubernetes.io/projected/e43e2506-d653-4464-9597-9579f4d7d3f6-kube-api-access-vn9tw\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.727806 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-config\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.831691 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.831783 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn9tw\" (UniqueName: \"kubernetes.io/projected/e43e2506-d653-4464-9597-9579f4d7d3f6-kube-api-access-vn9tw\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.831847 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-config\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.833038 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-config\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.833180 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.941403 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn9tw\" (UniqueName: \"kubernetes.io/projected/e43e2506-d653-4464-9597-9579f4d7d3f6-kube-api-access-vn9tw\") pod \"dnsmasq-dns-666b6646f7-jgxdf\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:16 crc kubenswrapper[4957]: I1206 06:00:16.952907 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.148817 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gdqct"] Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.180188 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d4s9x"] Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.189078 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.202484 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d4s9x"] Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.343473 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.343541 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-config\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.343577 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hp5q\" (UniqueName: \"kubernetes.io/projected/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-kube-api-access-9hp5q\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.364884 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jgxdf"] Dec 06 06:00:17 crc kubenswrapper[4957]: W1206 06:00:17.376331 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode43e2506_d653_4464_9597_9579f4d7d3f6.slice/crio-9a0c949978d2ec4f213dd013cc69e23798a576997547cd63900ea2a673091638 WatchSource:0}: Error finding container 9a0c949978d2ec4f213dd013cc69e23798a576997547cd63900ea2a673091638: Status 404 returned error can't find the container with id 9a0c949978d2ec4f213dd013cc69e23798a576997547cd63900ea2a673091638 Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.445860 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-config\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.445915 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hp5q\" (UniqueName: \"kubernetes.io/projected/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-kube-api-access-9hp5q\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.446347 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.447732 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-config\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.447739 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.467216 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hp5q\" (UniqueName: \"kubernetes.io/projected/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-kube-api-access-9hp5q\") pod \"dnsmasq-dns-57d769cc4f-d4s9x\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.519240 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.765326 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d4s9x"] Dec 06 06:00:17 crc kubenswrapper[4957]: W1206 06:00:17.778295 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4e09e52_89e8_45e3_a3b4_5e9e6a1b3b0c.slice/crio-7b92bc794658e48c2568d4527b4d5b0944a11f7936439e2afdbb44e99516da92 WatchSource:0}: Error finding container 7b92bc794658e48c2568d4527b4d5b0944a11f7936439e2afdbb44e99516da92: Status 404 returned error can't find the container with id 7b92bc794658e48c2568d4527b4d5b0944a11f7936439e2afdbb44e99516da92 Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.804079 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.806854 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.811396 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.811530 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.811432 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.811435 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.811479 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.812290 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ht8kt" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.812517 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.820514 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.952867 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.952953 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953003 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953033 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953080 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953112 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng6nc\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-kube-api-access-ng6nc\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953159 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953318 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953395 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.953455 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:17 crc kubenswrapper[4957]: I1206 06:00:17.955478 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-config-data\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056490 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056558 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056590 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-config-data\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056620 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056637 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056652 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056671 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056688 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056705 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng6nc\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-kube-api-access-ng6nc\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056726 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.056776 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.057017 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.057921 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.058066 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.059287 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-config-data\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.061466 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.061809 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.066486 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.067609 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.074814 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-server-conf\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.077639 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-pod-info\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.078938 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng6nc\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-kube-api-access-ng6nc\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.099617 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.146154 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.260672 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" event={"ID":"e43e2506-d653-4464-9597-9579f4d7d3f6","Type":"ContainerStarted","Data":"9a0c949978d2ec4f213dd013cc69e23798a576997547cd63900ea2a673091638"} Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.262748 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" event={"ID":"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c","Type":"ContainerStarted","Data":"7b92bc794658e48c2568d4527b4d5b0944a11f7936439e2afdbb44e99516da92"} Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.282370 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.285960 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.291453 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.291899 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.296105 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.302249 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.310091 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.310339 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.310527 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6vrsv" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.310646 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.460867 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461112 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461235 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461330 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461383 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461432 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fd5c\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-kube-api-access-8fd5c\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461455 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461647 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461753 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461799 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.461885 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563344 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563396 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563435 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563459 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563486 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563504 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563538 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563556 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563570 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563587 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fd5c\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-kube-api-access-8fd5c\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.563602 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.564323 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.564772 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.565107 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.565528 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.565982 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.566382 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.572434 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.572861 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.579429 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.581600 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.592470 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fd5c\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-kube-api-access-8fd5c\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.594894 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.652739 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:00:18 crc kubenswrapper[4957]: I1206 06:00:18.764194 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:00:18 crc kubenswrapper[4957]: W1206 06:00:18.810592 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf1ac671_ce48_4e6e_a1d6_4b7cbd079050.slice/crio-dc115a275194cedd37444c709e146a1930050199882fc44fe17fbed923390369 WatchSource:0}: Error finding container dc115a275194cedd37444c709e146a1930050199882fc44fe17fbed923390369: Status 404 returned error can't find the container with id dc115a275194cedd37444c709e146a1930050199882fc44fe17fbed923390369 Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.110762 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.280403 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af1ac671-ce48-4e6e-a1d6-4b7cbd079050","Type":"ContainerStarted","Data":"dc115a275194cedd37444c709e146a1930050199882fc44fe17fbed923390369"} Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.281781 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"91bdbaa2-3bc1-44b8-b326-8feb00dc400a","Type":"ContainerStarted","Data":"7f8238c9029a5057464044f35695da1be88f14aac5ec6a85c246052f119b465b"} Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.524848 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.526180 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.536189 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-wz9kf" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.536249 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.537954 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.538938 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.542120 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.543355 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687050 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687350 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687434 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-config-data-default\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687506 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687591 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-kolla-config\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687610 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72w4m\" (UniqueName: \"kubernetes.io/projected/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-kube-api-access-72w4m\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687690 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.687977 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.789424 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-kolla-config\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.790258 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72w4m\" (UniqueName: \"kubernetes.io/projected/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-kube-api-access-72w4m\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.790196 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-kolla-config\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.790363 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.794472 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.794591 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.794636 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.794709 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.794735 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-config-data-default\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.794772 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.796859 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.798095 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-config-data-default\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.799142 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.805936 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.806070 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.831220 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.840740 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72w4m\" (UniqueName: \"kubernetes.io/projected/3d87bf6f-a3d1-41ce-ba76-ed55aec8645e-kube-api-access-72w4m\") pod \"openstack-galera-0\" (UID: \"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e\") " pod="openstack/openstack-galera-0" Dec 06 06:00:19 crc kubenswrapper[4957]: I1206 06:00:19.856584 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.379327 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 06:00:20 crc kubenswrapper[4957]: W1206 06:00:20.438710 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d87bf6f_a3d1_41ce_ba76_ed55aec8645e.slice/crio-3f3f7ba0e69676b70041eb89d208bd6d269afc6a408fffda0b2f6c8a3365101d WatchSource:0}: Error finding container 3f3f7ba0e69676b70041eb89d208bd6d269afc6a408fffda0b2f6c8a3365101d: Status 404 returned error can't find the container with id 3f3f7ba0e69676b70041eb89d208bd6d269afc6a408fffda0b2f6c8a3365101d Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.919997 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.922984 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.934642 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.934761 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-lqd5n" Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.934785 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.935210 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 06 06:00:20 crc kubenswrapper[4957]: I1206 06:00:20.935510 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.126679 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.126801 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.126849 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.126884 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.126939 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.126974 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjfq7\" (UniqueName: \"kubernetes.io/projected/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-kube-api-access-fjfq7\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.127941 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.127996 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.225461 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.227720 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231360 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231432 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231458 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjfq7\" (UniqueName: \"kubernetes.io/projected/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-kube-api-access-fjfq7\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231496 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231514 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231606 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231666 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.231686 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.232243 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.232467 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-2xfn6" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.232993 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.232996 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.233058 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.236391 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.239913 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.241352 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.256153 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.260640 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.269026 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.274445 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjfq7\" (UniqueName: \"kubernetes.io/projected/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-kube-api-access-fjfq7\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.274493 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10af3da-4556-4f2f-bfb8-886a6a93bfc4-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c10af3da-4556-4f2f-bfb8-886a6a93bfc4\") " pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.332320 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528a2b84-7047-444d-997e-4a92e0b5dbf9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.332394 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/528a2b84-7047-444d-997e-4a92e0b5dbf9-kolla-config\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.332783 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbtxb\" (UniqueName: \"kubernetes.io/projected/528a2b84-7047-444d-997e-4a92e0b5dbf9-kube-api-access-pbtxb\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.332866 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/528a2b84-7047-444d-997e-4a92e0b5dbf9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.332907 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/528a2b84-7047-444d-997e-4a92e0b5dbf9-config-data\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.337142 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e","Type":"ContainerStarted","Data":"3f3f7ba0e69676b70041eb89d208bd6d269afc6a408fffda0b2f6c8a3365101d"} Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.434380 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbtxb\" (UniqueName: \"kubernetes.io/projected/528a2b84-7047-444d-997e-4a92e0b5dbf9-kube-api-access-pbtxb\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.434428 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/528a2b84-7047-444d-997e-4a92e0b5dbf9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.434462 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/528a2b84-7047-444d-997e-4a92e0b5dbf9-config-data\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.434519 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528a2b84-7047-444d-997e-4a92e0b5dbf9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.434749 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/528a2b84-7047-444d-997e-4a92e0b5dbf9-kolla-config\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.435453 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/528a2b84-7047-444d-997e-4a92e0b5dbf9-config-data\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.436717 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/528a2b84-7047-444d-997e-4a92e0b5dbf9-kolla-config\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.451124 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528a2b84-7047-444d-997e-4a92e0b5dbf9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.454071 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/528a2b84-7047-444d-997e-4a92e0b5dbf9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.456242 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbtxb\" (UniqueName: \"kubernetes.io/projected/528a2b84-7047-444d-997e-4a92e0b5dbf9-kube-api-access-pbtxb\") pod \"memcached-0\" (UID: \"528a2b84-7047-444d-997e-4a92e0b5dbf9\") " pod="openstack/memcached-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.568680 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 06:00:21 crc kubenswrapper[4957]: I1206 06:00:21.658486 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 06:00:22 crc kubenswrapper[4957]: I1206 06:00:22.161596 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 06:00:22 crc kubenswrapper[4957]: I1206 06:00:22.244001 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.006992 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.008803 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.010907 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-58spg" Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.029066 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.180860 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg7j5\" (UniqueName: \"kubernetes.io/projected/33344326-7541-47d7-8b16-0df124dce272-kube-api-access-tg7j5\") pod \"kube-state-metrics-0\" (UID: \"33344326-7541-47d7-8b16-0df124dce272\") " pod="openstack/kube-state-metrics-0" Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.282318 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg7j5\" (UniqueName: \"kubernetes.io/projected/33344326-7541-47d7-8b16-0df124dce272-kube-api-access-tg7j5\") pod \"kube-state-metrics-0\" (UID: \"33344326-7541-47d7-8b16-0df124dce272\") " pod="openstack/kube-state-metrics-0" Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.307516 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg7j5\" (UniqueName: \"kubernetes.io/projected/33344326-7541-47d7-8b16-0df124dce272-kube-api-access-tg7j5\") pod \"kube-state-metrics-0\" (UID: \"33344326-7541-47d7-8b16-0df124dce272\") " pod="openstack/kube-state-metrics-0" Dec 06 06:00:23 crc kubenswrapper[4957]: I1206 06:00:23.335688 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.280370 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pmd7t"] Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.281555 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.284373 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7fqxc" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.294548 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.294586 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.295582 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pmd7t"] Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.342340 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-qhgsm"] Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.345298 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.354897 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qhgsm"] Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.438619 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf193218-2a15-4cf5-8ba0-59dc939445ff-scripts\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.438666 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dfa83fc-3365-4d78-980f-23a2c71711f7-scripts\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.438699 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf193218-2a15-4cf5-8ba0-59dc939445ff-ovn-controller-tls-certs\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.438851 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf193218-2a15-4cf5-8ba0-59dc939445ff-combined-ca-bundle\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.438871 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-log-ovn\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439020 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-lib\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439060 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-run-ovn\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439135 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-run\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439168 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-etc-ovs\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439187 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wp99\" (UniqueName: \"kubernetes.io/projected/bf193218-2a15-4cf5-8ba0-59dc939445ff-kube-api-access-6wp99\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439218 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7px4t\" (UniqueName: \"kubernetes.io/projected/0dfa83fc-3365-4d78-980f-23a2c71711f7-kube-api-access-7px4t\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439241 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-run\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.439277 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-log\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540183 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf193218-2a15-4cf5-8ba0-59dc939445ff-combined-ca-bundle\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540237 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-log-ovn\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540282 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-lib\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540298 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-run-ovn\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540345 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-run\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540370 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-etc-ovs\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540386 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wp99\" (UniqueName: \"kubernetes.io/projected/bf193218-2a15-4cf5-8ba0-59dc939445ff-kube-api-access-6wp99\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540444 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7px4t\" (UniqueName: \"kubernetes.io/projected/0dfa83fc-3365-4d78-980f-23a2c71711f7-kube-api-access-7px4t\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540461 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-run\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540810 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-log\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540881 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf193218-2a15-4cf5-8ba0-59dc939445ff-scripts\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540903 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dfa83fc-3365-4d78-980f-23a2c71711f7-scripts\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.540932 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf193218-2a15-4cf5-8ba0-59dc939445ff-ovn-controller-tls-certs\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.541032 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-log-ovn\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.541187 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-run\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.541199 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-run\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.541263 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-etc-ovs\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.541264 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-log\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.541425 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf193218-2a15-4cf5-8ba0-59dc939445ff-var-run-ovn\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.543148 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dfa83fc-3365-4d78-980f-23a2c71711f7-scripts\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.543659 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0dfa83fc-3365-4d78-980f-23a2c71711f7-var-lib\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.548734 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf193218-2a15-4cf5-8ba0-59dc939445ff-combined-ca-bundle\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.559208 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf193218-2a15-4cf5-8ba0-59dc939445ff-ovn-controller-tls-certs\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.564329 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7px4t\" (UniqueName: \"kubernetes.io/projected/0dfa83fc-3365-4d78-980f-23a2c71711f7-kube-api-access-7px4t\") pod \"ovn-controller-ovs-qhgsm\" (UID: \"0dfa83fc-3365-4d78-980f-23a2c71711f7\") " pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.568472 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf193218-2a15-4cf5-8ba0-59dc939445ff-scripts\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.581437 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wp99\" (UniqueName: \"kubernetes.io/projected/bf193218-2a15-4cf5-8ba0-59dc939445ff-kube-api-access-6wp99\") pod \"ovn-controller-pmd7t\" (UID: \"bf193218-2a15-4cf5-8ba0-59dc939445ff\") " pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.616036 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:26 crc kubenswrapper[4957]: I1206 06:00:26.675561 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:27 crc kubenswrapper[4957]: W1206 06:00:27.318554 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc10af3da_4556_4f2f_bfb8_886a6a93bfc4.slice/crio-29237690a3c0d08e2937ba82a072293a2b8cf35bda160fe2b558d465ef96eeb1 WatchSource:0}: Error finding container 29237690a3c0d08e2937ba82a072293a2b8cf35bda160fe2b558d465ef96eeb1: Status 404 returned error can't find the container with id 29237690a3c0d08e2937ba82a072293a2b8cf35bda160fe2b558d465ef96eeb1 Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.385865 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c10af3da-4556-4f2f-bfb8-886a6a93bfc4","Type":"ContainerStarted","Data":"29237690a3c0d08e2937ba82a072293a2b8cf35bda160fe2b558d465ef96eeb1"} Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.389089 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"528a2b84-7047-444d-997e-4a92e0b5dbf9","Type":"ContainerStarted","Data":"d3a420da56f7152f6e0f2da7d4501cc288dd9fffb1cc68a666f4f83361cb2d6b"} Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.786050 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.790387 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.795818 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.795913 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.796220 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-9x2wb" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.796428 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.800850 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.823664 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.864726 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.864993 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6ec6d60-cadc-488c-bb21-514a7acbfc70-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.865048 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.865085 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.865132 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.865397 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d6ec6d60-cadc-488c-bb21-514a7acbfc70-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.865434 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n754c\" (UniqueName: \"kubernetes.io/projected/d6ec6d60-cadc-488c-bb21-514a7acbfc70-kube-api-access-n754c\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.865562 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec6d60-cadc-488c-bb21-514a7acbfc70-config\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.967790 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n754c\" (UniqueName: \"kubernetes.io/projected/d6ec6d60-cadc-488c-bb21-514a7acbfc70-kube-api-access-n754c\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.967941 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec6d60-cadc-488c-bb21-514a7acbfc70-config\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.967980 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.968012 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6ec6d60-cadc-488c-bb21-514a7acbfc70-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.968043 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.968078 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.968106 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.968133 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d6ec6d60-cadc-488c-bb21-514a7acbfc70-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.968715 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.969472 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d6ec6d60-cadc-488c-bb21-514a7acbfc70-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.969499 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6ec6d60-cadc-488c-bb21-514a7acbfc70-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.970406 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec6d60-cadc-488c-bb21-514a7acbfc70-config\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.973499 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.973519 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.974134 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ec6d60-cadc-488c-bb21-514a7acbfc70-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.983944 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n754c\" (UniqueName: \"kubernetes.io/projected/d6ec6d60-cadc-488c-bb21-514a7acbfc70-kube-api-access-n754c\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:27 crc kubenswrapper[4957]: I1206 06:00:27.989435 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d6ec6d60-cadc-488c-bb21-514a7acbfc70\") " pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:28 crc kubenswrapper[4957]: I1206 06:00:28.110568 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.572990 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.574283 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.577088 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2l82m" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.577237 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.577881 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.579968 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.585387 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690155 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6c710a2d-a836-4154-9b55-70e3d6ad61a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690221 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c710a2d-a836-4154-9b55-70e3d6ad61a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690263 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690407 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690482 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c710a2d-a836-4154-9b55-70e3d6ad61a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690550 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690588 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.690641 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb9hp\" (UniqueName: \"kubernetes.io/projected/6c710a2d-a836-4154-9b55-70e3d6ad61a1-kube-api-access-vb9hp\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792545 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb9hp\" (UniqueName: \"kubernetes.io/projected/6c710a2d-a836-4154-9b55-70e3d6ad61a1-kube-api-access-vb9hp\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792588 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6c710a2d-a836-4154-9b55-70e3d6ad61a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792624 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c710a2d-a836-4154-9b55-70e3d6ad61a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792657 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792699 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792732 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c710a2d-a836-4154-9b55-70e3d6ad61a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792762 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.792781 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.793249 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6c710a2d-a836-4154-9b55-70e3d6ad61a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.793299 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.793965 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c710a2d-a836-4154-9b55-70e3d6ad61a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.794375 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c710a2d-a836-4154-9b55-70e3d6ad61a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.797584 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.798040 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.805358 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c710a2d-a836-4154-9b55-70e3d6ad61a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.813277 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb9hp\" (UniqueName: \"kubernetes.io/projected/6c710a2d-a836-4154-9b55-70e3d6ad61a1-kube-api-access-vb9hp\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.826567 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6c710a2d-a836-4154-9b55-70e3d6ad61a1\") " pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:29 crc kubenswrapper[4957]: I1206 06:00:29.894599 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:38 crc kubenswrapper[4957]: E1206 06:00:38.053010 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 06 06:00:38 crc kubenswrapper[4957]: E1206 06:00:38.053529 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ng6nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(af1ac671-ce48-4e6e-a1d6-4b7cbd079050): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:38 crc kubenswrapper[4957]: E1206 06:00:38.054712 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" Dec 06 06:00:38 crc kubenswrapper[4957]: E1206 06:00:38.484332 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" Dec 06 06:00:39 crc kubenswrapper[4957]: E1206 06:00:39.920988 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 06 06:00:39 crc kubenswrapper[4957]: E1206 06:00:39.921483 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-72w4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(3d87bf6f-a3d1-41ce-ba76-ed55aec8645e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:39 crc kubenswrapper[4957]: E1206 06:00:39.923919 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="3d87bf6f-a3d1-41ce-ba76-ed55aec8645e" Dec 06 06:00:39 crc kubenswrapper[4957]: E1206 06:00:39.932380 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 06 06:00:39 crc kubenswrapper[4957]: E1206 06:00:39.932564 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8fd5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(91bdbaa2-3bc1-44b8-b326-8feb00dc400a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:39 crc kubenswrapper[4957]: E1206 06:00:39.933960 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" Dec 06 06:00:40 crc kubenswrapper[4957]: I1206 06:00:40.427978 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qhgsm"] Dec 06 06:00:40 crc kubenswrapper[4957]: E1206 06:00:40.497132 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="3d87bf6f-a3d1-41ce-ba76-ed55aec8645e" Dec 06 06:00:40 crc kubenswrapper[4957]: E1206 06:00:40.497646 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" Dec 06 06:00:44 crc kubenswrapper[4957]: E1206 06:00:44.355774 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 06 06:00:44 crc kubenswrapper[4957]: E1206 06:00:44.357971 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n68ch675h5b4h67chfdhfch547h558hffhf8h57fh59dh4h66h674h57dh685h57bh588hc4h559h54h67bh676h577h594h5ch5cfh678h685h667hb5q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pbtxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(528a2b84-7047-444d-997e-4a92e0b5dbf9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:44 crc kubenswrapper[4957]: E1206 06:00:44.363038 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="528a2b84-7047-444d-997e-4a92e0b5dbf9" Dec 06 06:00:44 crc kubenswrapper[4957]: I1206 06:00:44.527003 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qhgsm" event={"ID":"0dfa83fc-3365-4d78-980f-23a2c71711f7","Type":"ContainerStarted","Data":"1fa7462e3b84d9f76bad63c07760a995fcb9b6a78d862f0e76244b030ed8d2ad"} Dec 06 06:00:44 crc kubenswrapper[4957]: E1206 06:00:44.527974 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="528a2b84-7047-444d-997e-4a92e0b5dbf9" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.131469 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.131637 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bml9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-gdqct_openstack(a61521c5-8048-4bf1-be5c-6d84f0a915c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.133375 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" podUID="a61521c5-8048-4bf1-be5c-6d84f0a915c0" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.160537 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.160723 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8dc7j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-tlnjq_openstack(e8c48c03-e899-475e-aaad-5db38c19c26e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.162686 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" podUID="e8c48c03-e899-475e-aaad-5db38c19c26e" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.176481 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.176534 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.176687 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vn9tw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-jgxdf_openstack(e43e2506-d653-4464-9597-9579f4d7d3f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.176687 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9hp5q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-d4s9x_openstack(f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.178108 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" podUID="f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.178156 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" podUID="e43e2506-d653-4464-9597-9579f4d7d3f6" Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.536816 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c10af3da-4556-4f2f-bfb8-886a6a93bfc4","Type":"ContainerStarted","Data":"c3344f4245ebe288195cc58f49ff10386d7d3904c7ce4ea8e597fb85d09a398f"} Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.540549 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" podUID="e43e2506-d653-4464-9597-9579f4d7d3f6" Dec 06 06:00:45 crc kubenswrapper[4957]: E1206 06:00:45.540941 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" podUID="f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c" Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.610727 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.622532 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pmd7t"] Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.781205 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 06:00:45 crc kubenswrapper[4957]: W1206 06:00:45.797023 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6ec6d60_cadc_488c_bb21_514a7acbfc70.slice/crio-f16470c96e5f4993e9cd66e4c2072e48b1f4cccf65480a1d5706aa10d189a0da WatchSource:0}: Error finding container f16470c96e5f4993e9cd66e4c2072e48b1f4cccf65480a1d5706aa10d189a0da: Status 404 returned error can't find the container with id f16470c96e5f4993e9cd66e4c2072e48b1f4cccf65480a1d5706aa10d189a0da Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.874332 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.932324 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.976426 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bml9s\" (UniqueName: \"kubernetes.io/projected/a61521c5-8048-4bf1-be5c-6d84f0a915c0-kube-api-access-bml9s\") pod \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.976574 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-config\") pod \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.976629 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-dns-svc\") pod \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\" (UID: \"a61521c5-8048-4bf1-be5c-6d84f0a915c0\") " Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.977350 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-config" (OuterVolumeSpecName: "config") pod "a61521c5-8048-4bf1-be5c-6d84f0a915c0" (UID: "a61521c5-8048-4bf1-be5c-6d84f0a915c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.982750 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a61521c5-8048-4bf1-be5c-6d84f0a915c0-kube-api-access-bml9s" (OuterVolumeSpecName: "kube-api-access-bml9s") pod "a61521c5-8048-4bf1-be5c-6d84f0a915c0" (UID: "a61521c5-8048-4bf1-be5c-6d84f0a915c0"). InnerVolumeSpecName "kube-api-access-bml9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:00:45 crc kubenswrapper[4957]: I1206 06:00:45.983924 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a61521c5-8048-4bf1-be5c-6d84f0a915c0" (UID: "a61521c5-8048-4bf1-be5c-6d84f0a915c0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.078073 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dc7j\" (UniqueName: \"kubernetes.io/projected/e8c48c03-e899-475e-aaad-5db38c19c26e-kube-api-access-8dc7j\") pod \"e8c48c03-e899-475e-aaad-5db38c19c26e\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.078433 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c48c03-e899-475e-aaad-5db38c19c26e-config\") pod \"e8c48c03-e899-475e-aaad-5db38c19c26e\" (UID: \"e8c48c03-e899-475e-aaad-5db38c19c26e\") " Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.078784 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.078874 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61521c5-8048-4bf1-be5c-6d84f0a915c0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.078918 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8c48c03-e899-475e-aaad-5db38c19c26e-config" (OuterVolumeSpecName: "config") pod "e8c48c03-e899-475e-aaad-5db38c19c26e" (UID: "e8c48c03-e899-475e-aaad-5db38c19c26e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.078944 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bml9s\" (UniqueName: \"kubernetes.io/projected/a61521c5-8048-4bf1-be5c-6d84f0a915c0-kube-api-access-bml9s\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.081351 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c48c03-e899-475e-aaad-5db38c19c26e-kube-api-access-8dc7j" (OuterVolumeSpecName: "kube-api-access-8dc7j") pod "e8c48c03-e899-475e-aaad-5db38c19c26e" (UID: "e8c48c03-e899-475e-aaad-5db38c19c26e"). InnerVolumeSpecName "kube-api-access-8dc7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.191269 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8c48c03-e899-475e-aaad-5db38c19c26e-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.191331 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dc7j\" (UniqueName: \"kubernetes.io/projected/e8c48c03-e899-475e-aaad-5db38c19c26e-kube-api-access-8dc7j\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.544285 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" event={"ID":"e8c48c03-e899-475e-aaad-5db38c19c26e","Type":"ContainerDied","Data":"17cf687f6725ad87ba5a869a2011d2eb41340e7535d658e735e42093e1d5218e"} Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.544367 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tlnjq" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.549866 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" event={"ID":"a61521c5-8048-4bf1-be5c-6d84f0a915c0","Type":"ContainerDied","Data":"2fe24e8ae164a735ab666614e944efcc5c1832e28e4f8737db55a61a00c214a8"} Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.551677 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gdqct" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.553294 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"33344326-7541-47d7-8b16-0df124dce272","Type":"ContainerStarted","Data":"418f2509b76d912259234fad2184fcece35ecaeb4cf66e9bb338921883259f04"} Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.555091 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d6ec6d60-cadc-488c-bb21-514a7acbfc70","Type":"ContainerStarted","Data":"f16470c96e5f4993e9cd66e4c2072e48b1f4cccf65480a1d5706aa10d189a0da"} Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.557522 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t" event={"ID":"bf193218-2a15-4cf5-8ba0-59dc939445ff","Type":"ContainerStarted","Data":"d2b3cc6d018505da6f7854efb7040afd808d56c2963a0e5cf4830a9bd8daa8bf"} Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.595988 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.615031 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tlnjq"] Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.627545 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tlnjq"] Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.640120 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gdqct"] Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.645606 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gdqct"] Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.672710 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a61521c5-8048-4bf1-be5c-6d84f0a915c0" path="/var/lib/kubelet/pods/a61521c5-8048-4bf1-be5c-6d84f0a915c0/volumes" Dec 06 06:00:46 crc kubenswrapper[4957]: I1206 06:00:46.673165 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8c48c03-e899-475e-aaad-5db38c19c26e" path="/var/lib/kubelet/pods/e8c48c03-e899-475e-aaad-5db38c19c26e/volumes" Dec 06 06:00:48 crc kubenswrapper[4957]: W1206 06:00:48.558332 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c710a2d_a836_4154_9b55_70e3d6ad61a1.slice/crio-596f08b8cdb64f2beaf980dd51e50dc0d11dbfbd26e6b3c820cb9db21f4df485 WatchSource:0}: Error finding container 596f08b8cdb64f2beaf980dd51e50dc0d11dbfbd26e6b3c820cb9db21f4df485: Status 404 returned error can't find the container with id 596f08b8cdb64f2beaf980dd51e50dc0d11dbfbd26e6b3c820cb9db21f4df485 Dec 06 06:00:48 crc kubenswrapper[4957]: I1206 06:00:48.577019 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6c710a2d-a836-4154-9b55-70e3d6ad61a1","Type":"ContainerStarted","Data":"596f08b8cdb64f2beaf980dd51e50dc0d11dbfbd26e6b3c820cb9db21f4df485"} Dec 06 06:00:50 crc kubenswrapper[4957]: I1206 06:00:50.592956 4957 generic.go:334] "Generic (PLEG): container finished" podID="c10af3da-4556-4f2f-bfb8-886a6a93bfc4" containerID="c3344f4245ebe288195cc58f49ff10386d7d3904c7ce4ea8e597fb85d09a398f" exitCode=0 Dec 06 06:00:50 crc kubenswrapper[4957]: I1206 06:00:50.593059 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c10af3da-4556-4f2f-bfb8-886a6a93bfc4","Type":"ContainerDied","Data":"c3344f4245ebe288195cc58f49ff10386d7d3904c7ce4ea8e597fb85d09a398f"} Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.607173 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c10af3da-4556-4f2f-bfb8-886a6a93bfc4","Type":"ContainerStarted","Data":"51655b4f5a0d35761b401dca51675c7c1c503bd31cfa8a5cd73b07548edd4db5"} Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.609929 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d6ec6d60-cadc-488c-bb21-514a7acbfc70","Type":"ContainerStarted","Data":"e42c0ade64a70d282ad6153005f1f04976e57a99203c961fd5e85998aaa515f6"} Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.611903 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t" event={"ID":"bf193218-2a15-4cf5-8ba0-59dc939445ff","Type":"ContainerStarted","Data":"82c4cad284ae4144e48189fb8292ab25c5e090c5aead561121bc0005aab8cc12"} Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.612005 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-pmd7t" Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.613724 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6c710a2d-a836-4154-9b55-70e3d6ad61a1","Type":"ContainerStarted","Data":"e82b20cb8244aa6fd21c862816d74dca4323d783c3415b8802c03d4570b96a11"} Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.615589 4957 generic.go:334] "Generic (PLEG): container finished" podID="0dfa83fc-3365-4d78-980f-23a2c71711f7" containerID="aa010cb805cce1ec5b412d3417d8347053c9205606508a134a986d42a5b0b9de" exitCode=0 Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.615742 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qhgsm" event={"ID":"0dfa83fc-3365-4d78-980f-23a2c71711f7","Type":"ContainerDied","Data":"aa010cb805cce1ec5b412d3417d8347053c9205606508a134a986d42a5b0b9de"} Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.644096 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=14.80963368 podStartE2EDuration="32.644071005s" podCreationTimestamp="2025-12-06 06:00:19 +0000 UTC" firstStartedPulling="2025-12-06 06:00:27.328509955 +0000 UTC m=+1266.978777597" lastFinishedPulling="2025-12-06 06:00:45.16294729 +0000 UTC m=+1284.813214922" observedRunningTime="2025-12-06 06:00:51.639980194 +0000 UTC m=+1291.290247896" watchObservedRunningTime="2025-12-06 06:00:51.644071005 +0000 UTC m=+1291.294338647" Dec 06 06:00:51 crc kubenswrapper[4957]: I1206 06:00:51.679329 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-pmd7t" podStartSLOduration=21.845764506 podStartE2EDuration="25.679310761s" podCreationTimestamp="2025-12-06 06:00:26 +0000 UTC" firstStartedPulling="2025-12-06 06:00:45.625338742 +0000 UTC m=+1285.275606374" lastFinishedPulling="2025-12-06 06:00:49.458884997 +0000 UTC m=+1289.109152629" observedRunningTime="2025-12-06 06:00:51.675438694 +0000 UTC m=+1291.325706346" watchObservedRunningTime="2025-12-06 06:00:51.679310761 +0000 UTC m=+1291.329578393" Dec 06 06:00:52 crc kubenswrapper[4957]: I1206 06:00:52.626891 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af1ac671-ce48-4e6e-a1d6-4b7cbd079050","Type":"ContainerStarted","Data":"ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72"} Dec 06 06:00:52 crc kubenswrapper[4957]: I1206 06:00:52.641149 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qhgsm" event={"ID":"0dfa83fc-3365-4d78-980f-23a2c71711f7","Type":"ContainerStarted","Data":"25751472de325f827f76f8015764b219386f682ea20b30d4de251e1f0c723dff"} Dec 06 06:00:53 crc kubenswrapper[4957]: I1206 06:00:53.651739 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qhgsm" event={"ID":"0dfa83fc-3365-4d78-980f-23a2c71711f7","Type":"ContainerStarted","Data":"521c4df1210170786d138d4ceaf854b16b93fdef6e7a801a64afbfc80b9faa26"} Dec 06 06:00:53 crc kubenswrapper[4957]: I1206 06:00:53.652157 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:53 crc kubenswrapper[4957]: I1206 06:00:53.656526 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"33344326-7541-47d7-8b16-0df124dce272","Type":"ContainerStarted","Data":"2b5e3d35938d1d4b9b38309b7fec8bd84334a0f650e7a705fe5fc92b13443c2a"} Dec 06 06:00:53 crc kubenswrapper[4957]: I1206 06:00:53.657315 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 06:00:53 crc kubenswrapper[4957]: I1206 06:00:53.658885 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e","Type":"ContainerStarted","Data":"2cc571cf2b99d4841c8a16a06c42982346b2415ecf89865a92af47800213c111"} Dec 06 06:00:53 crc kubenswrapper[4957]: I1206 06:00:53.678902 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-qhgsm" podStartSLOduration=22.718637469 podStartE2EDuration="27.678882607s" podCreationTimestamp="2025-12-06 06:00:26 +0000 UTC" firstStartedPulling="2025-12-06 06:00:44.347308345 +0000 UTC m=+1283.997575977" lastFinishedPulling="2025-12-06 06:00:49.307553483 +0000 UTC m=+1288.957821115" observedRunningTime="2025-12-06 06:00:53.671151975 +0000 UTC m=+1293.321419627" watchObservedRunningTime="2025-12-06 06:00:53.678882607 +0000 UTC m=+1293.329150239" Dec 06 06:00:53 crc kubenswrapper[4957]: I1206 06:00:53.715683 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=25.0987315 podStartE2EDuration="31.715658434s" podCreationTimestamp="2025-12-06 06:00:22 +0000 UTC" firstStartedPulling="2025-12-06 06:00:45.632357744 +0000 UTC m=+1285.282625376" lastFinishedPulling="2025-12-06 06:00:52.249284678 +0000 UTC m=+1291.899552310" observedRunningTime="2025-12-06 06:00:53.712629141 +0000 UTC m=+1293.362896783" watchObservedRunningTime="2025-12-06 06:00:53.715658434 +0000 UTC m=+1293.365926066" Dec 06 06:00:54 crc kubenswrapper[4957]: I1206 06:00:54.670472 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:00:54 crc kubenswrapper[4957]: I1206 06:00:54.670853 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"91bdbaa2-3bc1-44b8-b326-8feb00dc400a","Type":"ContainerStarted","Data":"652119ddca1221ec06b591256f24db2bbd49ae927a7b0a92db1879b283119a69"} Dec 06 06:00:55 crc kubenswrapper[4957]: I1206 06:00:55.674444 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d6ec6d60-cadc-488c-bb21-514a7acbfc70","Type":"ContainerStarted","Data":"3b4043bfeed60e4d33ce366bde8eebcf2358cc24027a1e8cd11f47d71533def9"} Dec 06 06:00:55 crc kubenswrapper[4957]: I1206 06:00:55.677291 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6c710a2d-a836-4154-9b55-70e3d6ad61a1","Type":"ContainerStarted","Data":"a8b70d2413b8763a3cd852fca88e354ba96951f6af07e9c426ccab521dd671ba"} Dec 06 06:00:56 crc kubenswrapper[4957]: I1206 06:00:56.734894 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=22.784337214 podStartE2EDuration="28.734867711s" podCreationTimestamp="2025-12-06 06:00:28 +0000 UTC" firstStartedPulling="2025-12-06 06:00:48.563795247 +0000 UTC m=+1288.214062879" lastFinishedPulling="2025-12-06 06:00:54.514325744 +0000 UTC m=+1294.164593376" observedRunningTime="2025-12-06 06:00:56.713487145 +0000 UTC m=+1296.363754877" watchObservedRunningTime="2025-12-06 06:00:56.734867711 +0000 UTC m=+1296.385135343" Dec 06 06:00:56 crc kubenswrapper[4957]: I1206 06:00:56.743294 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=22.010179538 podStartE2EDuration="30.743262051s" podCreationTimestamp="2025-12-06 06:00:26 +0000 UTC" firstStartedPulling="2025-12-06 06:00:45.800669533 +0000 UTC m=+1285.450937155" lastFinishedPulling="2025-12-06 06:00:54.533752036 +0000 UTC m=+1294.184019668" observedRunningTime="2025-12-06 06:00:56.735853758 +0000 UTC m=+1296.386121390" watchObservedRunningTime="2025-12-06 06:00:56.743262051 +0000 UTC m=+1296.393529683" Dec 06 06:00:56 crc kubenswrapper[4957]: I1206 06:00:56.895741 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:56 crc kubenswrapper[4957]: I1206 06:00:56.934126 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:57 crc kubenswrapper[4957]: I1206 06:00:57.697629 4957 generic.go:334] "Generic (PLEG): container finished" podID="3d87bf6f-a3d1-41ce-ba76-ed55aec8645e" containerID="2cc571cf2b99d4841c8a16a06c42982346b2415ecf89865a92af47800213c111" exitCode=0 Dec 06 06:00:57 crc kubenswrapper[4957]: I1206 06:00:57.697713 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e","Type":"ContainerDied","Data":"2cc571cf2b99d4841c8a16a06c42982346b2415ecf89865a92af47800213c111"} Dec 06 06:00:57 crc kubenswrapper[4957]: I1206 06:00:57.701005 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"528a2b84-7047-444d-997e-4a92e0b5dbf9","Type":"ContainerStarted","Data":"3aa41ae4f8739d6708d2727511b620ddba61fa0a8403baf9310e147db94087ce"} Dec 06 06:00:57 crc kubenswrapper[4957]: I1206 06:00:57.701456 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:57 crc kubenswrapper[4957]: I1206 06:00:57.701496 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 06 06:00:57 crc kubenswrapper[4957]: I1206 06:00:57.761204 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=6.405776711 podStartE2EDuration="36.761181635s" podCreationTimestamp="2025-12-06 06:00:21 +0000 UTC" firstStartedPulling="2025-12-06 06:00:26.77615474 +0000 UTC m=+1266.426422392" lastFinishedPulling="2025-12-06 06:00:57.131559674 +0000 UTC m=+1296.781827316" observedRunningTime="2025-12-06 06:00:57.760555378 +0000 UTC m=+1297.410823030" watchObservedRunningTime="2025-12-06 06:00:57.761181635 +0000 UTC m=+1297.411449287" Dec 06 06:00:57 crc kubenswrapper[4957]: I1206 06:00:57.772499 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.044766 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jgxdf"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.089019 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-q7r5j"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.090927 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.094881 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x5dlh"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.096170 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.097626 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.103622 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.113339 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x5dlh"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.113423 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.114301 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.128719 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-q7r5j"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.171679 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.221870 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86fsn\" (UniqueName: \"kubernetes.io/projected/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-kube-api-access-86fsn\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.221926 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222027 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-config\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222052 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-ovn-rundir\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222081 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-ovs-rundir\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222130 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222154 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-config\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222209 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k77g\" (UniqueName: \"kubernetes.io/projected/baa42883-f7e2-475e-8350-2c0ea0a1e00c-kube-api-access-6k77g\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222243 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-combined-ca-bundle\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.222280 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325286 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325344 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-config\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325395 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k77g\" (UniqueName: \"kubernetes.io/projected/baa42883-f7e2-475e-8350-2c0ea0a1e00c-kube-api-access-6k77g\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325433 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-combined-ca-bundle\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325472 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325532 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86fsn\" (UniqueName: \"kubernetes.io/projected/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-kube-api-access-86fsn\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325561 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325619 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-config\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325643 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-ovn-rundir\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.325672 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-ovs-rundir\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.326051 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-ovs-rundir\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.326645 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-config\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.327016 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.328350 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.328911 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-config\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.329017 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-ovn-rundir\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.336478 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.336574 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-combined-ca-bundle\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.367998 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k77g\" (UniqueName: \"kubernetes.io/projected/baa42883-f7e2-475e-8350-2c0ea0a1e00c-kube-api-access-6k77g\") pod \"dnsmasq-dns-7f896c8c65-x5dlh\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.369407 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86fsn\" (UniqueName: \"kubernetes.io/projected/e5adfb1c-2a78-42af-af7e-419bf4d73bd4-kube-api-access-86fsn\") pod \"ovn-controller-metrics-q7r5j\" (UID: \"e5adfb1c-2a78-42af-af7e-419bf4d73bd4\") " pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.386343 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d4s9x"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.410210 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-q7r5j" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.431267 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.474349 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.520912 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8d9bm"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.534437 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.536223 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-dns-svc\") pod \"e43e2506-d653-4464-9597-9579f4d7d3f6\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.536360 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn9tw\" (UniqueName: \"kubernetes.io/projected/e43e2506-d653-4464-9597-9579f4d7d3f6-kube-api-access-vn9tw\") pod \"e43e2506-d653-4464-9597-9579f4d7d3f6\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.536488 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-config\") pod \"e43e2506-d653-4464-9597-9579f4d7d3f6\" (UID: \"e43e2506-d653-4464-9597-9579f4d7d3f6\") " Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.537159 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.537181 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e43e2506-d653-4464-9597-9579f4d7d3f6" (UID: "e43e2506-d653-4464-9597-9579f4d7d3f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.547036 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8d9bm"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.548153 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-config" (OuterVolumeSpecName: "config") pod "e43e2506-d653-4464-9597-9579f4d7d3f6" (UID: "e43e2506-d653-4464-9597-9579f4d7d3f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.559379 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43e2506-d653-4464-9597-9579f4d7d3f6-kube-api-access-vn9tw" (OuterVolumeSpecName: "kube-api-access-vn9tw") pod "e43e2506-d653-4464-9597-9579f4d7d3f6" (UID: "e43e2506-d653-4464-9597-9579f4d7d3f6"). InnerVolumeSpecName "kube-api-access-vn9tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638472 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638555 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-config\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638577 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwgxq\" (UniqueName: \"kubernetes.io/projected/60bc648f-d745-44d2-a574-3f15d4ca2764-kube-api-access-cwgxq\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638595 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638618 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638703 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn9tw\" (UniqueName: \"kubernetes.io/projected/e43e2506-d653-4464-9597-9579f4d7d3f6-kube-api-access-vn9tw\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638716 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.638725 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e43e2506-d653-4464-9597-9579f4d7d3f6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.718863 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3d87bf6f-a3d1-41ce-ba76-ed55aec8645e","Type":"ContainerStarted","Data":"3cf38122290781b1a574398d2a9ef14a19cb3a50f8c188f0555112e0ff7c91bb"} Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.721638 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" event={"ID":"e43e2506-d653-4464-9597-9579f4d7d3f6","Type":"ContainerDied","Data":"9a0c949978d2ec4f213dd013cc69e23798a576997547cd63900ea2a673091638"} Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.721878 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jgxdf" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.739783 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.740079 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-config\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.740116 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwgxq\" (UniqueName: \"kubernetes.io/projected/60bc648f-d745-44d2-a574-3f15d4ca2764-kube-api-access-cwgxq\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.740142 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.740176 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.741196 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.741649 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-config\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.742361 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.743110 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.767795 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwgxq\" (UniqueName: \"kubernetes.io/projected/60bc648f-d745-44d2-a574-3f15d4ca2764-kube-api-access-cwgxq\") pod \"dnsmasq-dns-86db49b7ff-8d9bm\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.780069 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371996.07473 podStartE2EDuration="40.780045696s" podCreationTimestamp="2025-12-06 06:00:18 +0000 UTC" firstStartedPulling="2025-12-06 06:00:20.441255026 +0000 UTC m=+1260.091522668" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:00:58.75175322 +0000 UTC m=+1298.402020872" watchObservedRunningTime="2025-12-06 06:00:58.780045696 +0000 UTC m=+1298.430313328" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.796148 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jgxdf"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.803757 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jgxdf"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.816971 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.817036 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.846894 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-dns-svc\") pod \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.847050 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-config\") pod \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.847181 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hp5q\" (UniqueName: \"kubernetes.io/projected/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-kube-api-access-9hp5q\") pod \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\" (UID: \"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c\") " Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.848632 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c" (UID: "f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.849014 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-config" (OuterVolumeSpecName: "config") pod "f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c" (UID: "f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.869514 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.873487 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-kube-api-access-9hp5q" (OuterVolumeSpecName: "kube-api-access-9hp5q") pod "f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c" (UID: "f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c"). InnerVolumeSpecName "kube-api-access-9hp5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.949180 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hp5q\" (UniqueName: \"kubernetes.io/projected/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-kube-api-access-9hp5q\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.949219 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.949228 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.957484 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.958716 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.965543 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-t84h7" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.965928 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.966143 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.966370 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 06 06:00:58 crc kubenswrapper[4957]: I1206 06:00:58.973041 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.050104 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.050161 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-scripts\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.050214 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.050271 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5xbx\" (UniqueName: \"kubernetes.io/projected/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-kube-api-access-q5xbx\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.050300 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-config\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.050316 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.050330 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.101979 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x5dlh"] Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.155636 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-config\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.155679 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.155700 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.155721 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.155753 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-scripts\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.155803 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.155873 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5xbx\" (UniqueName: \"kubernetes.io/projected/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-kube-api-access-q5xbx\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.156968 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-config\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.157221 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.160032 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-scripts\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.163325 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.167697 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.173887 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.175980 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5xbx\" (UniqueName: \"kubernetes.io/projected/7262d5d8-b26d-4871-a3c7-9ff1f2c4009f-kube-api-access-q5xbx\") pod \"ovn-northd-0\" (UID: \"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f\") " pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.267278 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-q7r5j"] Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.293371 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.430941 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8d9bm"] Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.751978 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-q7r5j" event={"ID":"e5adfb1c-2a78-42af-af7e-419bf4d73bd4","Type":"ContainerStarted","Data":"07b759cdf5e4a7b8a7bc83132ff3da20e0642f1602cf8063f047eed26177b435"} Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.752279 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-q7r5j" event={"ID":"e5adfb1c-2a78-42af-af7e-419bf4d73bd4","Type":"ContainerStarted","Data":"bb4ad5524d5a45626f9d9e5bf382be7fb995eaf136ed2573bcc34fe0fc459465"} Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.759466 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" event={"ID":"f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c","Type":"ContainerDied","Data":"7b92bc794658e48c2568d4527b4d5b0944a11f7936439e2afdbb44e99516da92"} Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.759650 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d4s9x" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.767275 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" event={"ID":"baa42883-f7e2-475e-8350-2c0ea0a1e00c","Type":"ContainerStarted","Data":"250dc6135b2c271293587935d36d9745e087ad84624bc966ada3fba0a46aab19"} Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.769090 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" event={"ID":"60bc648f-d745-44d2-a574-3f15d4ca2764","Type":"ContainerStarted","Data":"8248eb421bcf630efca4e00af3a1a155ed10bd243d592f4bde5936982ef55b36"} Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.773051 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.784595 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-q7r5j" podStartSLOduration=1.784571767 podStartE2EDuration="1.784571767s" podCreationTimestamp="2025-12-06 06:00:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:00:59.776780616 +0000 UTC m=+1299.427048248" watchObservedRunningTime="2025-12-06 06:00:59.784571767 +0000 UTC m=+1299.434839399" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.850803 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d4s9x"] Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.857277 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.857311 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 06:00:59 crc kubenswrapper[4957]: I1206 06:00:59.859579 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d4s9x"] Dec 06 06:01:00 crc kubenswrapper[4957]: I1206 06:01:00.680025 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43e2506-d653-4464-9597-9579f4d7d3f6" path="/var/lib/kubelet/pods/e43e2506-d653-4464-9597-9579f4d7d3f6/volumes" Dec 06 06:01:00 crc kubenswrapper[4957]: I1206 06:01:00.680849 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c" path="/var/lib/kubelet/pods/f4e09e52-89e8-45e3-a3b4-5e9e6a1b3b0c/volumes" Dec 06 06:01:00 crc kubenswrapper[4957]: I1206 06:01:00.782184 4957 generic.go:334] "Generic (PLEG): container finished" podID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerID="2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d" exitCode=0 Dec 06 06:01:00 crc kubenswrapper[4957]: I1206 06:01:00.782270 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" event={"ID":"baa42883-f7e2-475e-8350-2c0ea0a1e00c","Type":"ContainerDied","Data":"2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d"} Dec 06 06:01:00 crc kubenswrapper[4957]: I1206 06:01:00.784374 4957 generic.go:334] "Generic (PLEG): container finished" podID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerID="d847bbb963fb3ef0a55705a2d70684dd9c51c902352b6aaee253e1249fc7d84d" exitCode=0 Dec 06 06:01:00 crc kubenswrapper[4957]: I1206 06:01:00.784427 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" event={"ID":"60bc648f-d745-44d2-a574-3f15d4ca2764","Type":"ContainerDied","Data":"d847bbb963fb3ef0a55705a2d70684dd9c51c902352b6aaee253e1249fc7d84d"} Dec 06 06:01:00 crc kubenswrapper[4957]: I1206 06:01:00.787148 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f","Type":"ContainerStarted","Data":"5ce5b9e2fe568c695f4baa8792c8887c1cdc8060f5a99e8b25061dc52141dd38"} Dec 06 06:01:01 crc kubenswrapper[4957]: I1206 06:01:01.569039 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 06:01:01 crc kubenswrapper[4957]: I1206 06:01:01.569627 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 06:01:01 crc kubenswrapper[4957]: I1206 06:01:01.638404 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 06:01:01 crc kubenswrapper[4957]: I1206 06:01:01.891119 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 06:01:02 crc kubenswrapper[4957]: I1206 06:01:02.803449 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f","Type":"ContainerStarted","Data":"09d7696c724fcc4f709ce352ee1dd677ed50af15ee421f8dd9566960cdaf8e78"} Dec 06 06:01:02 crc kubenswrapper[4957]: I1206 06:01:02.805508 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" event={"ID":"baa42883-f7e2-475e-8350-2c0ea0a1e00c","Type":"ContainerStarted","Data":"3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b"} Dec 06 06:01:02 crc kubenswrapper[4957]: I1206 06:01:02.805759 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:01:02 crc kubenswrapper[4957]: I1206 06:01:02.810774 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" event={"ID":"60bc648f-d745-44d2-a574-3f15d4ca2764","Type":"ContainerStarted","Data":"566b84715ce462323808485d107b3ad9bac3628adf9ea66b30badb3a87e39e28"} Dec 06 06:01:02 crc kubenswrapper[4957]: I1206 06:01:02.810828 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:01:02 crc kubenswrapper[4957]: I1206 06:01:02.828503 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" podStartSLOduration=4.32826236 podStartE2EDuration="4.828489398s" podCreationTimestamp="2025-12-06 06:00:58 +0000 UTC" firstStartedPulling="2025-12-06 06:00:59.106924157 +0000 UTC m=+1298.757191799" lastFinishedPulling="2025-12-06 06:00:59.607151185 +0000 UTC m=+1299.257418837" observedRunningTime="2025-12-06 06:01:02.824416847 +0000 UTC m=+1302.474684489" watchObservedRunningTime="2025-12-06 06:01:02.828489398 +0000 UTC m=+1302.478757030" Dec 06 06:01:02 crc kubenswrapper[4957]: I1206 06:01:02.854684 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" podStartSLOduration=4.446864322 podStartE2EDuration="4.854667729s" podCreationTimestamp="2025-12-06 06:00:58 +0000 UTC" firstStartedPulling="2025-12-06 06:00:59.442929678 +0000 UTC m=+1299.093197310" lastFinishedPulling="2025-12-06 06:00:59.850733095 +0000 UTC m=+1299.501000717" observedRunningTime="2025-12-06 06:01:02.852410167 +0000 UTC m=+1302.502677819" watchObservedRunningTime="2025-12-06 06:01:02.854667729 +0000 UTC m=+1302.504935371" Dec 06 06:01:03 crc kubenswrapper[4957]: I1206 06:01:03.341150 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 06:01:03 crc kubenswrapper[4957]: I1206 06:01:03.818555 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7262d5d8-b26d-4871-a3c7-9ff1f2c4009f","Type":"ContainerStarted","Data":"19775c89bd3d332d72ef21acec2e9ac58893983ef4392ec20c255b66eac431a2"} Dec 06 06:01:03 crc kubenswrapper[4957]: I1206 06:01:03.839623 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.375537057 podStartE2EDuration="5.83960471s" podCreationTimestamp="2025-12-06 06:00:58 +0000 UTC" firstStartedPulling="2025-12-06 06:00:59.78467333 +0000 UTC m=+1299.434940972" lastFinishedPulling="2025-12-06 06:01:01.248740993 +0000 UTC m=+1300.899008625" observedRunningTime="2025-12-06 06:01:03.834059719 +0000 UTC m=+1303.484327361" watchObservedRunningTime="2025-12-06 06:01:03.83960471 +0000 UTC m=+1303.489872342" Dec 06 06:01:04 crc kubenswrapper[4957]: I1206 06:01:04.294288 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 06 06:01:04 crc kubenswrapper[4957]: I1206 06:01:04.309099 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 06:01:04 crc kubenswrapper[4957]: I1206 06:01:04.415185 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 06:01:06 crc kubenswrapper[4957]: I1206 06:01:06.659666 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 06 06:01:06 crc kubenswrapper[4957]: I1206 06:01:06.969174 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-46vj4"] Dec 06 06:01:06 crc kubenswrapper[4957]: I1206 06:01:06.970458 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46vj4" Dec 06 06:01:06 crc kubenswrapper[4957]: I1206 06:01:06.977820 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-d882-account-create-update-kwl95"] Dec 06 06:01:06 crc kubenswrapper[4957]: I1206 06:01:06.979025 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:06 crc kubenswrapper[4957]: I1206 06:01:06.981531 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 06 06:01:06 crc kubenswrapper[4957]: I1206 06:01:06.993749 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d882-account-create-update-kwl95"] Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.004919 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-46vj4"] Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.033549 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b3373d-2571-42a7-b265-ff1409d332a0-operator-scripts\") pod \"glance-db-create-46vj4\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " pod="openstack/glance-db-create-46vj4" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.033671 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn5d2\" (UniqueName: \"kubernetes.io/projected/c2b3373d-2571-42a7-b265-ff1409d332a0-kube-api-access-cn5d2\") pod \"glance-db-create-46vj4\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " pod="openstack/glance-db-create-46vj4" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.033719 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-444df\" (UniqueName: \"kubernetes.io/projected/c28de4c2-bcee-4913-9153-157e8e33dda0-kube-api-access-444df\") pod \"glance-d882-account-create-update-kwl95\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.034086 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c28de4c2-bcee-4913-9153-157e8e33dda0-operator-scripts\") pod \"glance-d882-account-create-update-kwl95\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.136066 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c28de4c2-bcee-4913-9153-157e8e33dda0-operator-scripts\") pod \"glance-d882-account-create-update-kwl95\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.136114 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b3373d-2571-42a7-b265-ff1409d332a0-operator-scripts\") pod \"glance-db-create-46vj4\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " pod="openstack/glance-db-create-46vj4" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.136185 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn5d2\" (UniqueName: \"kubernetes.io/projected/c2b3373d-2571-42a7-b265-ff1409d332a0-kube-api-access-cn5d2\") pod \"glance-db-create-46vj4\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " pod="openstack/glance-db-create-46vj4" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.136231 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-444df\" (UniqueName: \"kubernetes.io/projected/c28de4c2-bcee-4913-9153-157e8e33dda0-kube-api-access-444df\") pod \"glance-d882-account-create-update-kwl95\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.137264 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c28de4c2-bcee-4913-9153-157e8e33dda0-operator-scripts\") pod \"glance-d882-account-create-update-kwl95\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.137734 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b3373d-2571-42a7-b265-ff1409d332a0-operator-scripts\") pod \"glance-db-create-46vj4\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " pod="openstack/glance-db-create-46vj4" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.181536 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn5d2\" (UniqueName: \"kubernetes.io/projected/c2b3373d-2571-42a7-b265-ff1409d332a0-kube-api-access-cn5d2\") pod \"glance-db-create-46vj4\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " pod="openstack/glance-db-create-46vj4" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.184415 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-444df\" (UniqueName: \"kubernetes.io/projected/c28de4c2-bcee-4913-9153-157e8e33dda0-kube-api-access-444df\") pod \"glance-d882-account-create-update-kwl95\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.306231 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46vj4" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.317320 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.765391 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-46vj4"] Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.849329 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-46vj4" event={"ID":"c2b3373d-2571-42a7-b265-ff1409d332a0","Type":"ContainerStarted","Data":"2ea971d5f17b889e0caa4f6e168ea27c8551b9ca5ee7fb50a0071fcf43b56e50"} Dec 06 06:01:07 crc kubenswrapper[4957]: I1206 06:01:07.853585 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d882-account-create-update-kwl95"] Dec 06 06:01:07 crc kubenswrapper[4957]: W1206 06:01:07.858717 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc28de4c2_bcee_4913_9153_157e8e33dda0.slice/crio-73b05f5157c801bb7aaf81d59925154958c9ec4fad757a76149b2a822a10b3db WatchSource:0}: Error finding container 73b05f5157c801bb7aaf81d59925154958c9ec4fad757a76149b2a822a10b3db: Status 404 returned error can't find the container with id 73b05f5157c801bb7aaf81d59925154958c9ec4fad757a76149b2a822a10b3db Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.435133 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.859288 4957 generic.go:334] "Generic (PLEG): container finished" podID="c2b3373d-2571-42a7-b265-ff1409d332a0" containerID="32d5b55b0977f5c8dde43cb01e93b73ff4fa2f60bc6967aa36c11336e7389892" exitCode=0 Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.859393 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-46vj4" event={"ID":"c2b3373d-2571-42a7-b265-ff1409d332a0","Type":"ContainerDied","Data":"32d5b55b0977f5c8dde43cb01e93b73ff4fa2f60bc6967aa36c11336e7389892"} Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.861187 4957 generic.go:334] "Generic (PLEG): container finished" podID="c28de4c2-bcee-4913-9153-157e8e33dda0" containerID="04bbdddda2de1167150f8aa0ebcfef25ee7ec2f1cb8f7693a928eed16299a503" exitCode=0 Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.861242 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d882-account-create-update-kwl95" event={"ID":"c28de4c2-bcee-4913-9153-157e8e33dda0","Type":"ContainerDied","Data":"04bbdddda2de1167150f8aa0ebcfef25ee7ec2f1cb8f7693a928eed16299a503"} Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.861264 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d882-account-create-update-kwl95" event={"ID":"c28de4c2-bcee-4913-9153-157e8e33dda0","Type":"ContainerStarted","Data":"73b05f5157c801bb7aaf81d59925154958c9ec4fad757a76149b2a822a10b3db"} Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.871012 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.977536 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x5dlh"] Dec 06 06:01:08 crc kubenswrapper[4957]: I1206 06:01:08.977774 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" podUID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerName="dnsmasq-dns" containerID="cri-o://3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b" gracePeriod=10 Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.422209 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.482739 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-ovsdbserver-sb\") pod \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.482794 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-dns-svc\") pod \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.482905 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k77g\" (UniqueName: \"kubernetes.io/projected/baa42883-f7e2-475e-8350-2c0ea0a1e00c-kube-api-access-6k77g\") pod \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.482967 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-config\") pod \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\" (UID: \"baa42883-f7e2-475e-8350-2c0ea0a1e00c\") " Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.489093 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baa42883-f7e2-475e-8350-2c0ea0a1e00c-kube-api-access-6k77g" (OuterVolumeSpecName: "kube-api-access-6k77g") pod "baa42883-f7e2-475e-8350-2c0ea0a1e00c" (UID: "baa42883-f7e2-475e-8350-2c0ea0a1e00c"). InnerVolumeSpecName "kube-api-access-6k77g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.525453 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-config" (OuterVolumeSpecName: "config") pod "baa42883-f7e2-475e-8350-2c0ea0a1e00c" (UID: "baa42883-f7e2-475e-8350-2c0ea0a1e00c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.533790 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "baa42883-f7e2-475e-8350-2c0ea0a1e00c" (UID: "baa42883-f7e2-475e-8350-2c0ea0a1e00c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.536336 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "baa42883-f7e2-475e-8350-2c0ea0a1e00c" (UID: "baa42883-f7e2-475e-8350-2c0ea0a1e00c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.586071 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.586116 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.586139 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/baa42883-f7e2-475e-8350-2c0ea0a1e00c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.586157 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k77g\" (UniqueName: \"kubernetes.io/projected/baa42883-f7e2-475e-8350-2c0ea0a1e00c-kube-api-access-6k77g\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.871201 4957 generic.go:334] "Generic (PLEG): container finished" podID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerID="3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b" exitCode=0 Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.871294 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" event={"ID":"baa42883-f7e2-475e-8350-2c0ea0a1e00c","Type":"ContainerDied","Data":"3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b"} Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.871349 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" event={"ID":"baa42883-f7e2-475e-8350-2c0ea0a1e00c","Type":"ContainerDied","Data":"250dc6135b2c271293587935d36d9745e087ad84624bc966ada3fba0a46aab19"} Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.871377 4957 scope.go:117] "RemoveContainer" containerID="3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.871536 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x5dlh" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.904266 4957 scope.go:117] "RemoveContainer" containerID="2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.918782 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x5dlh"] Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.932296 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x5dlh"] Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.943981 4957 scope.go:117] "RemoveContainer" containerID="3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b" Dec 06 06:01:09 crc kubenswrapper[4957]: E1206 06:01:09.945058 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b\": container with ID starting with 3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b not found: ID does not exist" containerID="3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.945110 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b"} err="failed to get container status \"3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b\": rpc error: code = NotFound desc = could not find container \"3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b\": container with ID starting with 3c99e5f8b8ef15d72f6e3679af78ebed62fe1e0afe22c3cfc4648cdd77a5f88b not found: ID does not exist" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.945139 4957 scope.go:117] "RemoveContainer" containerID="2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d" Dec 06 06:01:09 crc kubenswrapper[4957]: E1206 06:01:09.945522 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d\": container with ID starting with 2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d not found: ID does not exist" containerID="2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d" Dec 06 06:01:09 crc kubenswrapper[4957]: I1206 06:01:09.945573 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d"} err="failed to get container status \"2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d\": rpc error: code = NotFound desc = could not find container \"2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d\": container with ID starting with 2dfd44a4e7454385e644d0f6d6dcf25c95a1dbc87771644344e03c5f4a5c9f3d not found: ID does not exist" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.220034 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.306353 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-444df\" (UniqueName: \"kubernetes.io/projected/c28de4c2-bcee-4913-9153-157e8e33dda0-kube-api-access-444df\") pod \"c28de4c2-bcee-4913-9153-157e8e33dda0\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.306502 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c28de4c2-bcee-4913-9153-157e8e33dda0-operator-scripts\") pod \"c28de4c2-bcee-4913-9153-157e8e33dda0\" (UID: \"c28de4c2-bcee-4913-9153-157e8e33dda0\") " Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.307255 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c28de4c2-bcee-4913-9153-157e8e33dda0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c28de4c2-bcee-4913-9153-157e8e33dda0" (UID: "c28de4c2-bcee-4913-9153-157e8e33dda0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.311698 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46vj4" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.323084 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c28de4c2-bcee-4913-9153-157e8e33dda0-kube-api-access-444df" (OuterVolumeSpecName: "kube-api-access-444df") pod "c28de4c2-bcee-4913-9153-157e8e33dda0" (UID: "c28de4c2-bcee-4913-9153-157e8e33dda0"). InnerVolumeSpecName "kube-api-access-444df". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.407461 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn5d2\" (UniqueName: \"kubernetes.io/projected/c2b3373d-2571-42a7-b265-ff1409d332a0-kube-api-access-cn5d2\") pod \"c2b3373d-2571-42a7-b265-ff1409d332a0\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.407569 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b3373d-2571-42a7-b265-ff1409d332a0-operator-scripts\") pod \"c2b3373d-2571-42a7-b265-ff1409d332a0\" (UID: \"c2b3373d-2571-42a7-b265-ff1409d332a0\") " Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.407929 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-444df\" (UniqueName: \"kubernetes.io/projected/c28de4c2-bcee-4913-9153-157e8e33dda0-kube-api-access-444df\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.407949 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c28de4c2-bcee-4913-9153-157e8e33dda0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.408151 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b3373d-2571-42a7-b265-ff1409d332a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2b3373d-2571-42a7-b265-ff1409d332a0" (UID: "c2b3373d-2571-42a7-b265-ff1409d332a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.412233 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b3373d-2571-42a7-b265-ff1409d332a0-kube-api-access-cn5d2" (OuterVolumeSpecName: "kube-api-access-cn5d2") pod "c2b3373d-2571-42a7-b265-ff1409d332a0" (UID: "c2b3373d-2571-42a7-b265-ff1409d332a0"). InnerVolumeSpecName "kube-api-access-cn5d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.509670 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b3373d-2571-42a7-b265-ff1409d332a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.509715 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn5d2\" (UniqueName: \"kubernetes.io/projected/c2b3373d-2571-42a7-b265-ff1409d332a0-kube-api-access-cn5d2\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.675021 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" path="/var/lib/kubelet/pods/baa42883-f7e2-475e-8350-2c0ea0a1e00c/volumes" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.879677 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d882-account-create-update-kwl95" event={"ID":"c28de4c2-bcee-4913-9153-157e8e33dda0","Type":"ContainerDied","Data":"73b05f5157c801bb7aaf81d59925154958c9ec4fad757a76149b2a822a10b3db"} Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.879733 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73b05f5157c801bb7aaf81d59925154958c9ec4fad757a76149b2a822a10b3db" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.879693 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d882-account-create-update-kwl95" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.882633 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-46vj4" event={"ID":"c2b3373d-2571-42a7-b265-ff1409d332a0","Type":"ContainerDied","Data":"2ea971d5f17b889e0caa4f6e168ea27c8551b9ca5ee7fb50a0071fcf43b56e50"} Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.882665 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ea971d5f17b889e0caa4f6e168ea27c8551b9ca5ee7fb50a0071fcf43b56e50" Dec 06 06:01:10 crc kubenswrapper[4957]: I1206 06:01:10.882711 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46vj4" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.138431 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-8wrvt"] Dec 06 06:01:11 crc kubenswrapper[4957]: E1206 06:01:11.138902 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c28de4c2-bcee-4913-9153-157e8e33dda0" containerName="mariadb-account-create-update" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.138937 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c28de4c2-bcee-4913-9153-157e8e33dda0" containerName="mariadb-account-create-update" Dec 06 06:01:11 crc kubenswrapper[4957]: E1206 06:01:11.138970 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerName="init" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.138981 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerName="init" Dec 06 06:01:11 crc kubenswrapper[4957]: E1206 06:01:11.139002 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerName="dnsmasq-dns" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.139013 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerName="dnsmasq-dns" Dec 06 06:01:11 crc kubenswrapper[4957]: E1206 06:01:11.139033 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b3373d-2571-42a7-b265-ff1409d332a0" containerName="mariadb-database-create" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.139044 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b3373d-2571-42a7-b265-ff1409d332a0" containerName="mariadb-database-create" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.139311 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c28de4c2-bcee-4913-9153-157e8e33dda0" containerName="mariadb-account-create-update" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.139340 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="baa42883-f7e2-475e-8350-2c0ea0a1e00c" containerName="dnsmasq-dns" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.139356 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b3373d-2571-42a7-b265-ff1409d332a0" containerName="mariadb-database-create" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.140168 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.146472 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8wrvt"] Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.221262 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3026135-ca82-4fab-a560-4cb0a64e227b-operator-scripts\") pod \"keystone-db-create-8wrvt\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.221382 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cqsq\" (UniqueName: \"kubernetes.io/projected/e3026135-ca82-4fab-a560-4cb0a64e227b-kube-api-access-4cqsq\") pod \"keystone-db-create-8wrvt\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.250605 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-03d3-account-create-update-gjflt"] Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.253793 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.258189 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.280901 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-03d3-account-create-update-gjflt"] Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.323262 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-operator-scripts\") pod \"keystone-03d3-account-create-update-gjflt\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.323397 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3026135-ca82-4fab-a560-4cb0a64e227b-operator-scripts\") pod \"keystone-db-create-8wrvt\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.323554 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cqsq\" (UniqueName: \"kubernetes.io/projected/e3026135-ca82-4fab-a560-4cb0a64e227b-kube-api-access-4cqsq\") pod \"keystone-db-create-8wrvt\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.323613 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8ccs\" (UniqueName: \"kubernetes.io/projected/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-kube-api-access-j8ccs\") pod \"keystone-03d3-account-create-update-gjflt\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.324551 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3026135-ca82-4fab-a560-4cb0a64e227b-operator-scripts\") pod \"keystone-db-create-8wrvt\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.344420 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cqsq\" (UniqueName: \"kubernetes.io/projected/e3026135-ca82-4fab-a560-4cb0a64e227b-kube-api-access-4cqsq\") pod \"keystone-db-create-8wrvt\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.425743 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-operator-scripts\") pod \"keystone-03d3-account-create-update-gjflt\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.425987 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8ccs\" (UniqueName: \"kubernetes.io/projected/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-kube-api-access-j8ccs\") pod \"keystone-03d3-account-create-update-gjflt\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.426603 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-operator-scripts\") pod \"keystone-03d3-account-create-update-gjflt\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.443464 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-hhr4x"] Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.444436 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.450414 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8ccs\" (UniqueName: \"kubernetes.io/projected/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-kube-api-access-j8ccs\") pod \"keystone-03d3-account-create-update-gjflt\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.461567 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.464747 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hhr4x"] Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.527643 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qblmn\" (UniqueName: \"kubernetes.io/projected/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-kube-api-access-qblmn\") pod \"placement-db-create-hhr4x\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.527726 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-operator-scripts\") pod \"placement-db-create-hhr4x\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.546651 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-60fa-account-create-update-mpw77"] Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.547602 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.552714 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.558037 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-60fa-account-create-update-mpw77"] Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.590661 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.629300 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-operator-scripts\") pod \"placement-db-create-hhr4x\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.629371 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnxd4\" (UniqueName: \"kubernetes.io/projected/58d697c5-d777-48f0-a82e-5b685102c543-kube-api-access-lnxd4\") pod \"placement-60fa-account-create-update-mpw77\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.629414 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d697c5-d777-48f0-a82e-5b685102c543-operator-scripts\") pod \"placement-60fa-account-create-update-mpw77\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.629440 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qblmn\" (UniqueName: \"kubernetes.io/projected/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-kube-api-access-qblmn\") pod \"placement-db-create-hhr4x\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.630436 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-operator-scripts\") pod \"placement-db-create-hhr4x\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.651081 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qblmn\" (UniqueName: \"kubernetes.io/projected/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-kube-api-access-qblmn\") pod \"placement-db-create-hhr4x\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.730774 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnxd4\" (UniqueName: \"kubernetes.io/projected/58d697c5-d777-48f0-a82e-5b685102c543-kube-api-access-lnxd4\") pod \"placement-60fa-account-create-update-mpw77\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.730854 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d697c5-d777-48f0-a82e-5b685102c543-operator-scripts\") pod \"placement-60fa-account-create-update-mpw77\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.731504 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d697c5-d777-48f0-a82e-5b685102c543-operator-scripts\") pod \"placement-60fa-account-create-update-mpw77\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.748927 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnxd4\" (UniqueName: \"kubernetes.io/projected/58d697c5-d777-48f0-a82e-5b685102c543-kube-api-access-lnxd4\") pod \"placement-60fa-account-create-update-mpw77\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.838553 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.862460 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:11 crc kubenswrapper[4957]: I1206 06:01:11.937756 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8wrvt"] Dec 06 06:01:11 crc kubenswrapper[4957]: W1206 06:01:11.983477 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3026135_ca82_4fab_a560_4cb0a64e227b.slice/crio-6d9a6ac699763343aec4d918a90d5defb549ce1f68dde21d8cfba15d9bf7b1a1 WatchSource:0}: Error finding container 6d9a6ac699763343aec4d918a90d5defb549ce1f68dde21d8cfba15d9bf7b1a1: Status 404 returned error can't find the container with id 6d9a6ac699763343aec4d918a90d5defb549ce1f68dde21d8cfba15d9bf7b1a1 Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.024854 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-03d3-account-create-update-gjflt"] Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.104282 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-scxvl"] Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.105610 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.109540 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.109584 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-prhgl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.127896 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-scxvl"] Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.138176 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62p5s\" (UniqueName: \"kubernetes.io/projected/98dc1a26-e2e0-4447-a841-8507984b8862-kube-api-access-62p5s\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.138297 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-combined-ca-bundle\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.138354 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-config-data\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.138435 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-db-sync-config-data\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.240850 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-combined-ca-bundle\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.240941 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-config-data\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.241005 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-db-sync-config-data\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.241055 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62p5s\" (UniqueName: \"kubernetes.io/projected/98dc1a26-e2e0-4447-a841-8507984b8862-kube-api-access-62p5s\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.245748 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-db-sync-config-data\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.245851 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-config-data\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.245919 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-combined-ca-bundle\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.263742 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62p5s\" (UniqueName: \"kubernetes.io/projected/98dc1a26-e2e0-4447-a841-8507984b8862-kube-api-access-62p5s\") pod \"glance-db-sync-scxvl\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.309545 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hhr4x"] Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.412069 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-60fa-account-create-update-mpw77"] Dec 06 06:01:12 crc kubenswrapper[4957]: W1206 06:01:12.421275 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58d697c5_d777_48f0_a82e_5b685102c543.slice/crio-e65a831605fa2151b46ea1fc815e3f346b4c6b1b404d3c21a5394d3a0fb3b545 WatchSource:0}: Error finding container e65a831605fa2151b46ea1fc815e3f346b4c6b1b404d3c21a5394d3a0fb3b545: Status 404 returned error can't find the container with id e65a831605fa2151b46ea1fc815e3f346b4c6b1b404d3c21a5394d3a0fb3b545 Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.441641 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.898472 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-03d3-account-create-update-gjflt" event={"ID":"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9","Type":"ContainerStarted","Data":"bccad0d088bd0fccd396fad3450b9fbb1609c862785242a8c80af425fcac1ceb"} Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.898774 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-03d3-account-create-update-gjflt" event={"ID":"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9","Type":"ContainerStarted","Data":"550d911229bb10ace720aac110e896246246be4c1f38cff570d827d2b8792911"} Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.901059 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-60fa-account-create-update-mpw77" event={"ID":"58d697c5-d777-48f0-a82e-5b685102c543","Type":"ContainerStarted","Data":"e65a831605fa2151b46ea1fc815e3f346b4c6b1b404d3c21a5394d3a0fb3b545"} Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.904156 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8wrvt" event={"ID":"e3026135-ca82-4fab-a560-4cb0a64e227b","Type":"ContainerStarted","Data":"5f6e0499693b01d434bc35403cb87cd843b292a977574015ac0be27aa71db586"} Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.904397 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8wrvt" event={"ID":"e3026135-ca82-4fab-a560-4cb0a64e227b","Type":"ContainerStarted","Data":"6d9a6ac699763343aec4d918a90d5defb549ce1f68dde21d8cfba15d9bf7b1a1"} Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.906403 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hhr4x" event={"ID":"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb","Type":"ContainerStarted","Data":"75a019cdc82c607fa83bf943181412c9d8f6d0e4017f4b9dbfcd287cb4bac2d7"} Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.925250 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-03d3-account-create-update-gjflt" podStartSLOduration=1.925225988 podStartE2EDuration="1.925225988s" podCreationTimestamp="2025-12-06 06:01:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:12.917771345 +0000 UTC m=+1312.568038977" watchObservedRunningTime="2025-12-06 06:01:12.925225988 +0000 UTC m=+1312.575493640" Dec 06 06:01:12 crc kubenswrapper[4957]: I1206 06:01:12.939619 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-8wrvt" podStartSLOduration=1.939601738 podStartE2EDuration="1.939601738s" podCreationTimestamp="2025-12-06 06:01:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:12.93781907 +0000 UTC m=+1312.588086712" watchObservedRunningTime="2025-12-06 06:01:12.939601738 +0000 UTC m=+1312.589869380" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.042555 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-scxvl"] Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.323165 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-5mpr7"] Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.326883 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.428921 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5mpr7"] Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.495649 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.495761 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnc76\" (UniqueName: \"kubernetes.io/projected/e440d460-8e15-44ea-9500-476aba81d2cf-kube-api-access-cnc76\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.495797 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.495823 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-config\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.495863 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-dns-svc\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.597480 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.597610 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnc76\" (UniqueName: \"kubernetes.io/projected/e440d460-8e15-44ea-9500-476aba81d2cf-kube-api-access-cnc76\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.597656 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.597691 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-config\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.597718 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-dns-svc\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.598399 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.598548 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-dns-svc\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.599144 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.599243 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-config\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.632248 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnc76\" (UniqueName: \"kubernetes.io/projected/e440d460-8e15-44ea-9500-476aba81d2cf-kube-api-access-cnc76\") pod \"dnsmasq-dns-698758b865-5mpr7\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.672447 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.930293 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hhr4x" event={"ID":"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb","Type":"ContainerStarted","Data":"6719e8e343360b869dca20c4e5d038934e54fd5acf1a13276a5c1cde052816f4"} Dec 06 06:01:13 crc kubenswrapper[4957]: I1206 06:01:13.932578 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scxvl" event={"ID":"98dc1a26-e2e0-4447-a841-8507984b8862","Type":"ContainerStarted","Data":"2b85c5463bf37621af49baf44276a82296892748e42fc854359eb816d08ef976"} Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.147862 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5mpr7"] Dec 06 06:01:14 crc kubenswrapper[4957]: W1206 06:01:14.152270 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode440d460_8e15_44ea_9500_476aba81d2cf.slice/crio-c6b0f032c4b671d7c8c7d3e150dd2bda9bd65e858c554f77158812f4bbc2ac72 WatchSource:0}: Error finding container c6b0f032c4b671d7c8c7d3e150dd2bda9bd65e858c554f77158812f4bbc2ac72: Status 404 returned error can't find the container with id c6b0f032c4b671d7c8c7d3e150dd2bda9bd65e858c554f77158812f4bbc2ac72 Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.358600 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.513333 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.519107 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.521192 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.521496 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.521584 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-dd5x4" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.521619 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.542066 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.714639 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/72358664-e14f-482d-bd35-c807922b525f-cache\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.715029 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.715177 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.715342 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/72358664-e14f-482d-bd35-c807922b525f-lock\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.715429 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvfn8\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-kube-api-access-mvfn8\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.817971 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/72358664-e14f-482d-bd35-c807922b525f-cache\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.818086 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.818184 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.818273 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/72358664-e14f-482d-bd35-c807922b525f-lock\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.818314 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvfn8\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-kube-api-access-mvfn8\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.818501 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.818662 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/72358664-e14f-482d-bd35-c807922b525f-cache\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: E1206 06:01:14.818797 4957 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:01:14 crc kubenswrapper[4957]: E1206 06:01:14.818822 4957 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:01:14 crc kubenswrapper[4957]: E1206 06:01:14.818969 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift podName:72358664-e14f-482d-bd35-c807922b525f nodeName:}" failed. No retries permitted until 2025-12-06 06:01:15.31894602 +0000 UTC m=+1314.969213882 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift") pod "swift-storage-0" (UID: "72358664-e14f-482d-bd35-c807922b525f") : configmap "swift-ring-files" not found Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.819136 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/72358664-e14f-482d-bd35-c807922b525f-lock\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.839715 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvfn8\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-kube-api-access-mvfn8\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.843119 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.944341 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5mpr7" event={"ID":"e440d460-8e15-44ea-9500-476aba81d2cf","Type":"ContainerStarted","Data":"c6b0f032c4b671d7c8c7d3e150dd2bda9bd65e858c554f77158812f4bbc2ac72"} Dec 06 06:01:14 crc kubenswrapper[4957]: I1206 06:01:14.946434 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-60fa-account-create-update-mpw77" event={"ID":"58d697c5-d777-48f0-a82e-5b685102c543","Type":"ContainerStarted","Data":"c6ab871bb415b2afe7e5583240f10e5ccc07a9811e34299e559d34a4ef6dd403"} Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.080478 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wm72g"] Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.081682 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.084576 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.085271 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.085318 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.096108 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wm72g"] Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.223990 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-dispersionconf\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.224049 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-scripts\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.224123 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9s9m\" (UniqueName: \"kubernetes.io/projected/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-kube-api-access-s9s9m\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.224171 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-swiftconf\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.224459 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-combined-ca-bundle\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.224498 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-etc-swift\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.224705 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-ring-data-devices\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.325698 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.326047 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-combined-ca-bundle\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.326070 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-etc-swift\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: E1206 06:01:15.325850 4957 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.326119 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-ring-data-devices\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: E1206 06:01:15.326124 4957 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.326155 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-dispersionconf\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: E1206 06:01:15.326169 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift podName:72358664-e14f-482d-bd35-c807922b525f nodeName:}" failed. No retries permitted until 2025-12-06 06:01:16.326153886 +0000 UTC m=+1315.976421518 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift") pod "swift-storage-0" (UID: "72358664-e14f-482d-bd35-c807922b525f") : configmap "swift-ring-files" not found Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.326184 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-scripts\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.326208 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9s9m\" (UniqueName: \"kubernetes.io/projected/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-kube-api-access-s9s9m\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.326242 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-swiftconf\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.327068 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-etc-swift\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.327716 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-ring-data-devices\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.328069 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-scripts\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.330969 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-combined-ca-bundle\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.331296 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-swiftconf\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.332284 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-dispersionconf\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.350630 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9s9m\" (UniqueName: \"kubernetes.io/projected/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-kube-api-access-s9s9m\") pod \"swift-ring-rebalance-wm72g\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.398007 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.825860 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wm72g"] Dec 06 06:01:15 crc kubenswrapper[4957]: W1206 06:01:15.830259 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43f9d83d_effb_4bc6_8550_6e4c32adaf8d.slice/crio-70102655e753f1c7abeb95e867f4a39266b02730cb07eeebb65de8e450e6d123 WatchSource:0}: Error finding container 70102655e753f1c7abeb95e867f4a39266b02730cb07eeebb65de8e450e6d123: Status 404 returned error can't find the container with id 70102655e753f1c7abeb95e867f4a39266b02730cb07eeebb65de8e450e6d123 Dec 06 06:01:15 crc kubenswrapper[4957]: I1206 06:01:15.955737 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wm72g" event={"ID":"43f9d83d-effb-4bc6-8550-6e4c32adaf8d","Type":"ContainerStarted","Data":"70102655e753f1c7abeb95e867f4a39266b02730cb07eeebb65de8e450e6d123"} Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.341269 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:16 crc kubenswrapper[4957]: E1206 06:01:16.341490 4957 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:01:16 crc kubenswrapper[4957]: E1206 06:01:16.341518 4957 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:01:16 crc kubenswrapper[4957]: E1206 06:01:16.341579 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift podName:72358664-e14f-482d-bd35-c807922b525f nodeName:}" failed. No retries permitted until 2025-12-06 06:01:18.341561623 +0000 UTC m=+1317.991829255 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift") pod "swift-storage-0" (UID: "72358664-e14f-482d-bd35-c807922b525f") : configmap "swift-ring-files" not found Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.964423 4957 generic.go:334] "Generic (PLEG): container finished" podID="e440d460-8e15-44ea-9500-476aba81d2cf" containerID="bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9" exitCode=0 Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.964516 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5mpr7" event={"ID":"e440d460-8e15-44ea-9500-476aba81d2cf","Type":"ContainerDied","Data":"bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9"} Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.966247 4957 generic.go:334] "Generic (PLEG): container finished" podID="e3026135-ca82-4fab-a560-4cb0a64e227b" containerID="5f6e0499693b01d434bc35403cb87cd843b292a977574015ac0be27aa71db586" exitCode=0 Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.966302 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8wrvt" event={"ID":"e3026135-ca82-4fab-a560-4cb0a64e227b","Type":"ContainerDied","Data":"5f6e0499693b01d434bc35403cb87cd843b292a977574015ac0be27aa71db586"} Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.969524 4957 generic.go:334] "Generic (PLEG): container finished" podID="b4fe16c0-afe2-4b62-b62b-108ee7b0dddb" containerID="6719e8e343360b869dca20c4e5d038934e54fd5acf1a13276a5c1cde052816f4" exitCode=0 Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.969600 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hhr4x" event={"ID":"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb","Type":"ContainerDied","Data":"6719e8e343360b869dca20c4e5d038934e54fd5acf1a13276a5c1cde052816f4"} Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.971941 4957 generic.go:334] "Generic (PLEG): container finished" podID="16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9" containerID="bccad0d088bd0fccd396fad3450b9fbb1609c862785242a8c80af425fcac1ceb" exitCode=0 Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.972053 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-03d3-account-create-update-gjflt" event={"ID":"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9","Type":"ContainerDied","Data":"bccad0d088bd0fccd396fad3450b9fbb1609c862785242a8c80af425fcac1ceb"} Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.980311 4957 generic.go:334] "Generic (PLEG): container finished" podID="58d697c5-d777-48f0-a82e-5b685102c543" containerID="c6ab871bb415b2afe7e5583240f10e5ccc07a9811e34299e559d34a4ef6dd403" exitCode=0 Dec 06 06:01:16 crc kubenswrapper[4957]: I1206 06:01:16.980535 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-60fa-account-create-update-mpw77" event={"ID":"58d697c5-d777-48f0-a82e-5b685102c543","Type":"ContainerDied","Data":"c6ab871bb415b2afe7e5583240f10e5ccc07a9811e34299e559d34a4ef6dd403"} Dec 06 06:01:17 crc kubenswrapper[4957]: I1206 06:01:17.993493 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5mpr7" event={"ID":"e440d460-8e15-44ea-9500-476aba81d2cf","Type":"ContainerStarted","Data":"d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075"} Dec 06 06:01:17 crc kubenswrapper[4957]: I1206 06:01:17.994017 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:18 crc kubenswrapper[4957]: I1206 06:01:18.021011 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-5mpr7" podStartSLOduration=5.020988375 podStartE2EDuration="5.020988375s" podCreationTimestamp="2025-12-06 06:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:18.011890179 +0000 UTC m=+1317.662157891" watchObservedRunningTime="2025-12-06 06:01:18.020988375 +0000 UTC m=+1317.671256017" Dec 06 06:01:18 crc kubenswrapper[4957]: I1206 06:01:18.380076 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:18 crc kubenswrapper[4957]: E1206 06:01:18.380422 4957 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:01:18 crc kubenswrapper[4957]: E1206 06:01:18.380473 4957 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:01:18 crc kubenswrapper[4957]: E1206 06:01:18.380570 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift podName:72358664-e14f-482d-bd35-c807922b525f nodeName:}" failed. No retries permitted until 2025-12-06 06:01:22.38053808 +0000 UTC m=+1322.030805712 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift") pod "swift-storage-0" (UID: "72358664-e14f-482d-bd35-c807922b525f") : configmap "swift-ring-files" not found Dec 06 06:01:20 crc kubenswrapper[4957]: I1206 06:01:20.857103 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:20 crc kubenswrapper[4957]: I1206 06:01:20.871291 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:20 crc kubenswrapper[4957]: I1206 06:01:20.916745 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:20 crc kubenswrapper[4957]: I1206 06:01:20.929893 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.015870 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hhr4x" event={"ID":"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb","Type":"ContainerDied","Data":"75a019cdc82c607fa83bf943181412c9d8f6d0e4017f4b9dbfcd287cb4bac2d7"} Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.015923 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75a019cdc82c607fa83bf943181412c9d8f6d0e4017f4b9dbfcd287cb4bac2d7" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.015880 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hhr4x" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.017457 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-03d3-account-create-update-gjflt" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.017458 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-03d3-account-create-update-gjflt" event={"ID":"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9","Type":"ContainerDied","Data":"550d911229bb10ace720aac110e896246246be4c1f38cff570d827d2b8792911"} Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.017496 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="550d911229bb10ace720aac110e896246246be4c1f38cff570d827d2b8792911" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.018758 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-60fa-account-create-update-mpw77" event={"ID":"58d697c5-d777-48f0-a82e-5b685102c543","Type":"ContainerDied","Data":"e65a831605fa2151b46ea1fc815e3f346b4c6b1b404d3c21a5394d3a0fb3b545"} Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.018779 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e65a831605fa2151b46ea1fc815e3f346b4c6b1b404d3c21a5394d3a0fb3b545" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.018825 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-60fa-account-create-update-mpw77" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.020264 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wm72g" event={"ID":"43f9d83d-effb-4bc6-8550-6e4c32adaf8d","Type":"ContainerStarted","Data":"9ea4253102d6a4e66d238e81bbf9ae97d8a66639657cee20d7d0cba62631d10e"} Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.023986 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8wrvt" event={"ID":"e3026135-ca82-4fab-a560-4cb0a64e227b","Type":"ContainerDied","Data":"6d9a6ac699763343aec4d918a90d5defb549ce1f68dde21d8cfba15d9bf7b1a1"} Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.024025 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d9a6ac699763343aec4d918a90d5defb549ce1f68dde21d8cfba15d9bf7b1a1" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.023997 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8wrvt" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.028622 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8ccs\" (UniqueName: \"kubernetes.io/projected/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-kube-api-access-j8ccs\") pod \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.028677 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnxd4\" (UniqueName: \"kubernetes.io/projected/58d697c5-d777-48f0-a82e-5b685102c543-kube-api-access-lnxd4\") pod \"58d697c5-d777-48f0-a82e-5b685102c543\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.028778 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cqsq\" (UniqueName: \"kubernetes.io/projected/e3026135-ca82-4fab-a560-4cb0a64e227b-kube-api-access-4cqsq\") pod \"e3026135-ca82-4fab-a560-4cb0a64e227b\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.028810 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d697c5-d777-48f0-a82e-5b685102c543-operator-scripts\") pod \"58d697c5-d777-48f0-a82e-5b685102c543\" (UID: \"58d697c5-d777-48f0-a82e-5b685102c543\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.028870 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3026135-ca82-4fab-a560-4cb0a64e227b-operator-scripts\") pod \"e3026135-ca82-4fab-a560-4cb0a64e227b\" (UID: \"e3026135-ca82-4fab-a560-4cb0a64e227b\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.028900 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-operator-scripts\") pod \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\" (UID: \"16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.029915 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9" (UID: "16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.030394 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58d697c5-d777-48f0-a82e-5b685102c543-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "58d697c5-d777-48f0-a82e-5b685102c543" (UID: "58d697c5-d777-48f0-a82e-5b685102c543"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.030503 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3026135-ca82-4fab-a560-4cb0a64e227b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e3026135-ca82-4fab-a560-4cb0a64e227b" (UID: "e3026135-ca82-4fab-a560-4cb0a64e227b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.037049 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3026135-ca82-4fab-a560-4cb0a64e227b-kube-api-access-4cqsq" (OuterVolumeSpecName: "kube-api-access-4cqsq") pod "e3026135-ca82-4fab-a560-4cb0a64e227b" (UID: "e3026135-ca82-4fab-a560-4cb0a64e227b"). InnerVolumeSpecName "kube-api-access-4cqsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.037187 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-kube-api-access-j8ccs" (OuterVolumeSpecName: "kube-api-access-j8ccs") pod "16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9" (UID: "16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9"). InnerVolumeSpecName "kube-api-access-j8ccs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.038196 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d697c5-d777-48f0-a82e-5b685102c543-kube-api-access-lnxd4" (OuterVolumeSpecName: "kube-api-access-lnxd4") pod "58d697c5-d777-48f0-a82e-5b685102c543" (UID: "58d697c5-d777-48f0-a82e-5b685102c543"). InnerVolumeSpecName "kube-api-access-lnxd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.042671 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-wm72g" podStartSLOduration=1.199576897 podStartE2EDuration="6.042648612s" podCreationTimestamp="2025-12-06 06:01:15 +0000 UTC" firstStartedPulling="2025-12-06 06:01:15.834482041 +0000 UTC m=+1315.484749673" lastFinishedPulling="2025-12-06 06:01:20.677553746 +0000 UTC m=+1320.327821388" observedRunningTime="2025-12-06 06:01:21.037318507 +0000 UTC m=+1320.687586149" watchObservedRunningTime="2025-12-06 06:01:21.042648612 +0000 UTC m=+1320.692916244" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.129795 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-operator-scripts\") pod \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.129977 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qblmn\" (UniqueName: \"kubernetes.io/projected/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-kube-api-access-qblmn\") pod \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\" (UID: \"b4fe16c0-afe2-4b62-b62b-108ee7b0dddb\") " Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.130316 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4fe16c0-afe2-4b62-b62b-108ee7b0dddb" (UID: "b4fe16c0-afe2-4b62-b62b-108ee7b0dddb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.133066 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnxd4\" (UniqueName: \"kubernetes.io/projected/58d697c5-d777-48f0-a82e-5b685102c543-kube-api-access-lnxd4\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.133121 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cqsq\" (UniqueName: \"kubernetes.io/projected/e3026135-ca82-4fab-a560-4cb0a64e227b-kube-api-access-4cqsq\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.133142 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d697c5-d777-48f0-a82e-5b685102c543-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.133160 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3026135-ca82-4fab-a560-4cb0a64e227b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.133178 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.133196 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.133213 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8ccs\" (UniqueName: \"kubernetes.io/projected/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9-kube-api-access-j8ccs\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.134429 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-kube-api-access-qblmn" (OuterVolumeSpecName: "kube-api-access-qblmn") pod "b4fe16c0-afe2-4b62-b62b-108ee7b0dddb" (UID: "b4fe16c0-afe2-4b62-b62b-108ee7b0dddb"). InnerVolumeSpecName "kube-api-access-qblmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.234415 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qblmn\" (UniqueName: \"kubernetes.io/projected/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb-kube-api-access-qblmn\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.669997 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-pmd7t" podUID="bf193218-2a15-4cf5-8ba0-59dc939445ff" containerName="ovn-controller" probeResult="failure" output=< Dec 06 06:01:21 crc kubenswrapper[4957]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 06:01:21 crc kubenswrapper[4957]: > Dec 06 06:01:21 crc kubenswrapper[4957]: I1206 06:01:21.730583 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:01:22 crc kubenswrapper[4957]: I1206 06:01:22.460774 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:22 crc kubenswrapper[4957]: E1206 06:01:22.460999 4957 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 06:01:22 crc kubenswrapper[4957]: E1206 06:01:22.461272 4957 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 06:01:22 crc kubenswrapper[4957]: E1206 06:01:22.461338 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift podName:72358664-e14f-482d-bd35-c807922b525f nodeName:}" failed. No retries permitted until 2025-12-06 06:01:30.461320222 +0000 UTC m=+1330.111587854 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift") pod "swift-storage-0" (UID: "72358664-e14f-482d-bd35-c807922b525f") : configmap "swift-ring-files" not found Dec 06 06:01:23 crc kubenswrapper[4957]: I1206 06:01:23.674275 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:23 crc kubenswrapper[4957]: I1206 06:01:23.741385 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8d9bm"] Dec 06 06:01:23 crc kubenswrapper[4957]: I1206 06:01:23.741605 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerName="dnsmasq-dns" containerID="cri-o://566b84715ce462323808485d107b3ad9bac3628adf9ea66b30badb3a87e39e28" gracePeriod=10 Dec 06 06:01:23 crc kubenswrapper[4957]: I1206 06:01:23.871424 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: connect: connection refused" Dec 06 06:01:24 crc kubenswrapper[4957]: I1206 06:01:24.048080 4957 generic.go:334] "Generic (PLEG): container finished" podID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerID="ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72" exitCode=0 Dec 06 06:01:24 crc kubenswrapper[4957]: I1206 06:01:24.048102 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af1ac671-ce48-4e6e-a1d6-4b7cbd079050","Type":"ContainerDied","Data":"ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72"} Dec 06 06:01:24 crc kubenswrapper[4957]: I1206 06:01:24.054275 4957 generic.go:334] "Generic (PLEG): container finished" podID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerID="566b84715ce462323808485d107b3ad9bac3628adf9ea66b30badb3a87e39e28" exitCode=0 Dec 06 06:01:24 crc kubenswrapper[4957]: I1206 06:01:24.056233 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" event={"ID":"60bc648f-d745-44d2-a574-3f15d4ca2764","Type":"ContainerDied","Data":"566b84715ce462323808485d107b3ad9bac3628adf9ea66b30badb3a87e39e28"} Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.070429 4957 generic.go:334] "Generic (PLEG): container finished" podID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerID="652119ddca1221ec06b591256f24db2bbd49ae927a7b0a92db1879b283119a69" exitCode=0 Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.070491 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"91bdbaa2-3bc1-44b8-b326-8feb00dc400a","Type":"ContainerDied","Data":"652119ddca1221ec06b591256f24db2bbd49ae927a7b0a92db1879b283119a69"} Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.678082 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-pmd7t" podUID="bf193218-2a15-4cf5-8ba0-59dc939445ff" containerName="ovn-controller" probeResult="failure" output=< Dec 06 06:01:26 crc kubenswrapper[4957]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 06:01:26 crc kubenswrapper[4957]: > Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.718341 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qhgsm" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922174 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pmd7t-config-7jkpf"] Dec 06 06:01:26 crc kubenswrapper[4957]: E1206 06:01:26.922604 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9" containerName="mariadb-account-create-update" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922625 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9" containerName="mariadb-account-create-update" Dec 06 06:01:26 crc kubenswrapper[4957]: E1206 06:01:26.922643 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3026135-ca82-4fab-a560-4cb0a64e227b" containerName="mariadb-database-create" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922652 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3026135-ca82-4fab-a560-4cb0a64e227b" containerName="mariadb-database-create" Dec 06 06:01:26 crc kubenswrapper[4957]: E1206 06:01:26.922667 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4fe16c0-afe2-4b62-b62b-108ee7b0dddb" containerName="mariadb-database-create" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922675 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4fe16c0-afe2-4b62-b62b-108ee7b0dddb" containerName="mariadb-database-create" Dec 06 06:01:26 crc kubenswrapper[4957]: E1206 06:01:26.922694 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d697c5-d777-48f0-a82e-5b685102c543" containerName="mariadb-account-create-update" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922701 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d697c5-d777-48f0-a82e-5b685102c543" containerName="mariadb-account-create-update" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922929 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3026135-ca82-4fab-a560-4cb0a64e227b" containerName="mariadb-database-create" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922954 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9" containerName="mariadb-account-create-update" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922969 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4fe16c0-afe2-4b62-b62b-108ee7b0dddb" containerName="mariadb-database-create" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.922980 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d697c5-d777-48f0-a82e-5b685102c543" containerName="mariadb-account-create-update" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.923609 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.925889 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 06:01:26 crc kubenswrapper[4957]: I1206 06:01:26.938947 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pmd7t-config-7jkpf"] Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.045474 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run-ovn\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.045597 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-scripts\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.045676 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.045816 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-log-ovn\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.046436 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-additional-scripts\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.046506 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrdh2\" (UniqueName: \"kubernetes.io/projected/b64d0f4d-f513-42de-8512-1104a22adb15-kube-api-access-hrdh2\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.147853 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrdh2\" (UniqueName: \"kubernetes.io/projected/b64d0f4d-f513-42de-8512-1104a22adb15-kube-api-access-hrdh2\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.147967 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run-ovn\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148018 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-scripts\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148057 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148093 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-log-ovn\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148112 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-additional-scripts\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148343 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run-ovn\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148360 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148360 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-log-ovn\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.148922 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-additional-scripts\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.150185 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-scripts\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.164419 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrdh2\" (UniqueName: \"kubernetes.io/projected/b64d0f4d-f513-42de-8512-1104a22adb15-kube-api-access-hrdh2\") pod \"ovn-controller-pmd7t-config-7jkpf\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:27 crc kubenswrapper[4957]: I1206 06:01:27.259473 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.720436 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.893167 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-sb\") pod \"60bc648f-d745-44d2-a574-3f15d4ca2764\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.893249 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-dns-svc\") pod \"60bc648f-d745-44d2-a574-3f15d4ca2764\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.893282 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-config\") pod \"60bc648f-d745-44d2-a574-3f15d4ca2764\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.893315 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwgxq\" (UniqueName: \"kubernetes.io/projected/60bc648f-d745-44d2-a574-3f15d4ca2764-kube-api-access-cwgxq\") pod \"60bc648f-d745-44d2-a574-3f15d4ca2764\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.893413 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-nb\") pod \"60bc648f-d745-44d2-a574-3f15d4ca2764\" (UID: \"60bc648f-d745-44d2-a574-3f15d4ca2764\") " Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.906654 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60bc648f-d745-44d2-a574-3f15d4ca2764-kube-api-access-cwgxq" (OuterVolumeSpecName: "kube-api-access-cwgxq") pod "60bc648f-d745-44d2-a574-3f15d4ca2764" (UID: "60bc648f-d745-44d2-a574-3f15d4ca2764"). InnerVolumeSpecName "kube-api-access-cwgxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.942180 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "60bc648f-d745-44d2-a574-3f15d4ca2764" (UID: "60bc648f-d745-44d2-a574-3f15d4ca2764"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.952569 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pmd7t-config-7jkpf"] Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.954273 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-config" (OuterVolumeSpecName: "config") pod "60bc648f-d745-44d2-a574-3f15d4ca2764" (UID: "60bc648f-d745-44d2-a574-3f15d4ca2764"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.955368 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60bc648f-d745-44d2-a574-3f15d4ca2764" (UID: "60bc648f-d745-44d2-a574-3f15d4ca2764"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.994525 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60bc648f-d745-44d2-a574-3f15d4ca2764" (UID: "60bc648f-d745-44d2-a574-3f15d4ca2764"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.998889 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.998925 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.998934 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.998943 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwgxq\" (UniqueName: \"kubernetes.io/projected/60bc648f-d745-44d2-a574-3f15d4ca2764-kube-api-access-cwgxq\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:28 crc kubenswrapper[4957]: I1206 06:01:28.998953 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60bc648f-d745-44d2-a574-3f15d4ca2764-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.095678 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"91bdbaa2-3bc1-44b8-b326-8feb00dc400a","Type":"ContainerStarted","Data":"cbbbd9c5a84f18728f60f9cb9c74c08a957e41c603243d9a8d1717df9c96dfe4"} Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.097257 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.099978 4957 generic.go:334] "Generic (PLEG): container finished" podID="43f9d83d-effb-4bc6-8550-6e4c32adaf8d" containerID="9ea4253102d6a4e66d238e81bbf9ae97d8a66639657cee20d7d0cba62631d10e" exitCode=0 Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.100043 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wm72g" event={"ID":"43f9d83d-effb-4bc6-8550-6e4c32adaf8d","Type":"ContainerDied","Data":"9ea4253102d6a4e66d238e81bbf9ae97d8a66639657cee20d7d0cba62631d10e"} Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.103235 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" event={"ID":"60bc648f-d745-44d2-a574-3f15d4ca2764","Type":"ContainerDied","Data":"8248eb421bcf630efca4e00af3a1a155ed10bd243d592f4bde5936982ef55b36"} Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.103380 4957 scope.go:117] "RemoveContainer" containerID="566b84715ce462323808485d107b3ad9bac3628adf9ea66b30badb3a87e39e28" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.103591 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-8d9bm" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.107261 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t-config-7jkpf" event={"ID":"b64d0f4d-f513-42de-8512-1104a22adb15","Type":"ContainerStarted","Data":"40311453b2a991956cd666b81174b0f235ab50253ecbfb43f216557f5d57c968"} Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.113612 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af1ac671-ce48-4e6e-a1d6-4b7cbd079050","Type":"ContainerStarted","Data":"b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df"} Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.114465 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.140360 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371964.714571 podStartE2EDuration="1m12.140205435s" podCreationTimestamp="2025-12-06 06:00:17 +0000 UTC" firstStartedPulling="2025-12-06 06:00:19.117424174 +0000 UTC m=+1258.767691806" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:29.138028086 +0000 UTC m=+1328.788295738" watchObservedRunningTime="2025-12-06 06:01:29.140205435 +0000 UTC m=+1328.790473087" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.167245 4957 scope.go:117] "RemoveContainer" containerID="d847bbb963fb3ef0a55705a2d70684dd9c51c902352b6aaee253e1249fc7d84d" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.211097 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.221426523 podStartE2EDuration="1m13.211071939s" podCreationTimestamp="2025-12-06 06:00:16 +0000 UTC" firstStartedPulling="2025-12-06 06:00:18.841694594 +0000 UTC m=+1258.491962226" lastFinishedPulling="2025-12-06 06:00:50.83134001 +0000 UTC m=+1290.481607642" observedRunningTime="2025-12-06 06:01:29.189079953 +0000 UTC m=+1328.839347595" watchObservedRunningTime="2025-12-06 06:01:29.211071939 +0000 UTC m=+1328.861339571" Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.241591 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8d9bm"] Dec 06 06:01:29 crc kubenswrapper[4957]: I1206 06:01:29.261164 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-8d9bm"] Dec 06 06:01:29 crc kubenswrapper[4957]: E1206 06:01:29.849046 4957 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb64d0f4d_f513_42de_8512_1104a22adb15.slice/crio-923df90cc3dd9392e73c7f54746e648ac69f436f1ab2bb03747077e0bf10a116.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.148474 4957 generic.go:334] "Generic (PLEG): container finished" podID="b64d0f4d-f513-42de-8512-1104a22adb15" containerID="923df90cc3dd9392e73c7f54746e648ac69f436f1ab2bb03747077e0bf10a116" exitCode=0 Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.148744 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t-config-7jkpf" event={"ID":"b64d0f4d-f513-42de-8512-1104a22adb15","Type":"ContainerDied","Data":"923df90cc3dd9392e73c7f54746e648ac69f436f1ab2bb03747077e0bf10a116"} Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.152803 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scxvl" event={"ID":"98dc1a26-e2e0-4447-a841-8507984b8862","Type":"ContainerStarted","Data":"6303e4d02345bcf49a6dfb7dea4862406d20b2007dfb8ddbcd167a45067e42e2"} Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.195965 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-scxvl" podStartSLOduration=2.667930198 podStartE2EDuration="18.195939618s" podCreationTimestamp="2025-12-06 06:01:12 +0000 UTC" firstStartedPulling="2025-12-06 06:01:13.042047401 +0000 UTC m=+1312.692315033" lastFinishedPulling="2025-12-06 06:01:28.570056821 +0000 UTC m=+1328.220324453" observedRunningTime="2025-12-06 06:01:30.188180477 +0000 UTC m=+1329.838448149" watchObservedRunningTime="2025-12-06 06:01:30.195939618 +0000 UTC m=+1329.846207290" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.524513 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.542633 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.553472 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/72358664-e14f-482d-bd35-c807922b525f-etc-swift\") pod \"swift-storage-0\" (UID: \"72358664-e14f-482d-bd35-c807922b525f\") " pod="openstack/swift-storage-0" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.643667 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9s9m\" (UniqueName: \"kubernetes.io/projected/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-kube-api-access-s9s9m\") pod \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.643755 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-scripts\") pod \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.643787 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-dispersionconf\") pod \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.644082 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-ring-data-devices\") pod \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.644110 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-swiftconf\") pod \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.644139 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-etc-swift\") pod \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.644251 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-combined-ca-bundle\") pod \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\" (UID: \"43f9d83d-effb-4bc6-8550-6e4c32adaf8d\") " Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.645938 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "43f9d83d-effb-4bc6-8550-6e4c32adaf8d" (UID: "43f9d83d-effb-4bc6-8550-6e4c32adaf8d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.647634 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "43f9d83d-effb-4bc6-8550-6e4c32adaf8d" (UID: "43f9d83d-effb-4bc6-8550-6e4c32adaf8d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.649160 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-kube-api-access-s9s9m" (OuterVolumeSpecName: "kube-api-access-s9s9m") pod "43f9d83d-effb-4bc6-8550-6e4c32adaf8d" (UID: "43f9d83d-effb-4bc6-8550-6e4c32adaf8d"). InnerVolumeSpecName "kube-api-access-s9s9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.654216 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "43f9d83d-effb-4bc6-8550-6e4c32adaf8d" (UID: "43f9d83d-effb-4bc6-8550-6e4c32adaf8d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.676985 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" path="/var/lib/kubelet/pods/60bc648f-d745-44d2-a574-3f15d4ca2764/volumes" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.681017 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43f9d83d-effb-4bc6-8550-6e4c32adaf8d" (UID: "43f9d83d-effb-4bc6-8550-6e4c32adaf8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.707027 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "43f9d83d-effb-4bc6-8550-6e4c32adaf8d" (UID: "43f9d83d-effb-4bc6-8550-6e4c32adaf8d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.710323 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-scripts" (OuterVolumeSpecName: "scripts") pod "43f9d83d-effb-4bc6-8550-6e4c32adaf8d" (UID: "43f9d83d-effb-4bc6-8550-6e4c32adaf8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.736881 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.746568 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.746607 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9s9m\" (UniqueName: \"kubernetes.io/projected/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-kube-api-access-s9s9m\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.746622 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.746636 4957 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.746648 4957 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.746662 4957 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:30 crc kubenswrapper[4957]: I1206 06:01:30.746673 4957 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/43f9d83d-effb-4bc6-8550-6e4c32adaf8d-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.167624 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wm72g" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.167680 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wm72g" event={"ID":"43f9d83d-effb-4bc6-8550-6e4c32adaf8d","Type":"ContainerDied","Data":"70102655e753f1c7abeb95e867f4a39266b02730cb07eeebb65de8e450e6d123"} Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.168132 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70102655e753f1c7abeb95e867f4a39266b02730cb07eeebb65de8e450e6d123" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.325845 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.508997 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.661628 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-scripts\") pod \"b64d0f4d-f513-42de-8512-1104a22adb15\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.661741 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-additional-scripts\") pod \"b64d0f4d-f513-42de-8512-1104a22adb15\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.661817 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-log-ovn\") pod \"b64d0f4d-f513-42de-8512-1104a22adb15\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.661885 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run\") pod \"b64d0f4d-f513-42de-8512-1104a22adb15\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.661936 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run-ovn\") pod \"b64d0f4d-f513-42de-8512-1104a22adb15\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.662011 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrdh2\" (UniqueName: \"kubernetes.io/projected/b64d0f4d-f513-42de-8512-1104a22adb15-kube-api-access-hrdh2\") pod \"b64d0f4d-f513-42de-8512-1104a22adb15\" (UID: \"b64d0f4d-f513-42de-8512-1104a22adb15\") " Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.662267 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run" (OuterVolumeSpecName: "var-run") pod "b64d0f4d-f513-42de-8512-1104a22adb15" (UID: "b64d0f4d-f513-42de-8512-1104a22adb15"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.662314 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b64d0f4d-f513-42de-8512-1104a22adb15" (UID: "b64d0f4d-f513-42de-8512-1104a22adb15"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.662337 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b64d0f4d-f513-42de-8512-1104a22adb15" (UID: "b64d0f4d-f513-42de-8512-1104a22adb15"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.662738 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b64d0f4d-f513-42de-8512-1104a22adb15" (UID: "b64d0f4d-f513-42de-8512-1104a22adb15"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.663082 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-scripts" (OuterVolumeSpecName: "scripts") pod "b64d0f4d-f513-42de-8512-1104a22adb15" (UID: "b64d0f4d-f513-42de-8512-1104a22adb15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.667643 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-pmd7t" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.669078 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b64d0f4d-f513-42de-8512-1104a22adb15-kube-api-access-hrdh2" (OuterVolumeSpecName: "kube-api-access-hrdh2") pod "b64d0f4d-f513-42de-8512-1104a22adb15" (UID: "b64d0f4d-f513-42de-8512-1104a22adb15"). InnerVolumeSpecName "kube-api-access-hrdh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.763941 4957 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.763988 4957 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.763996 4957 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.764005 4957 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b64d0f4d-f513-42de-8512-1104a22adb15-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.764118 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrdh2\" (UniqueName: \"kubernetes.io/projected/b64d0f4d-f513-42de-8512-1104a22adb15-kube-api-access-hrdh2\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:31 crc kubenswrapper[4957]: I1206 06:01:31.764239 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b64d0f4d-f513-42de-8512-1104a22adb15-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.188725 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"86fcdc43c6d22050ea3c24c01ff8823a671ccd7e0ac99cb9bf9ccfc70bbfb5eb"} Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.191731 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t-config-7jkpf" event={"ID":"b64d0f4d-f513-42de-8512-1104a22adb15","Type":"ContainerDied","Data":"40311453b2a991956cd666b81174b0f235ab50253ecbfb43f216557f5d57c968"} Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.191759 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40311453b2a991956cd666b81174b0f235ab50253ecbfb43f216557f5d57c968" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.191800 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-7jkpf" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.750659 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-pmd7t-config-7jkpf"] Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.756623 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-pmd7t-config-7jkpf"] Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.866997 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pmd7t-config-bq5dj"] Dec 06 06:01:32 crc kubenswrapper[4957]: E1206 06:01:32.867457 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f9d83d-effb-4bc6-8550-6e4c32adaf8d" containerName="swift-ring-rebalance" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.867472 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f9d83d-effb-4bc6-8550-6e4c32adaf8d" containerName="swift-ring-rebalance" Dec 06 06:01:32 crc kubenswrapper[4957]: E1206 06:01:32.867491 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerName="init" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.867498 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerName="init" Dec 06 06:01:32 crc kubenswrapper[4957]: E1206 06:01:32.867541 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerName="dnsmasq-dns" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.867549 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerName="dnsmasq-dns" Dec 06 06:01:32 crc kubenswrapper[4957]: E1206 06:01:32.867559 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64d0f4d-f513-42de-8512-1104a22adb15" containerName="ovn-config" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.867567 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64d0f4d-f513-42de-8512-1104a22adb15" containerName="ovn-config" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.867794 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bc648f-d745-44d2-a574-3f15d4ca2764" containerName="dnsmasq-dns" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.867808 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b64d0f4d-f513-42de-8512-1104a22adb15" containerName="ovn-config" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.867825 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="43f9d83d-effb-4bc6-8550-6e4c32adaf8d" containerName="swift-ring-rebalance" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.868518 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.872481 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.873796 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pmd7t-config-bq5dj"] Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.980655 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-additional-scripts\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.980727 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq755\" (UniqueName: \"kubernetes.io/projected/06d0255b-0a8d-45f2-acc6-27985b748cad-kube-api-access-mq755\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.980927 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.980973 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-log-ovn\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.981046 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run-ovn\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:32 crc kubenswrapper[4957]: I1206 06:01:32.981081 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-scripts\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.082727 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq755\" (UniqueName: \"kubernetes.io/projected/06d0255b-0a8d-45f2-acc6-27985b748cad-kube-api-access-mq755\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.082796 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.082820 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-log-ovn\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.082867 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run-ovn\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.082889 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-scripts\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.082927 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-additional-scripts\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.083455 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-log-ovn\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.083455 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.083491 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run-ovn\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.085425 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-scripts\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.085750 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-additional-scripts\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.104432 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq755\" (UniqueName: \"kubernetes.io/projected/06d0255b-0a8d-45f2-acc6-27985b748cad-kube-api-access-mq755\") pod \"ovn-controller-pmd7t-config-bq5dj\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.198437 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.201759 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"5a18d118e8a2cffb49f97d663d3a21c63b6a533502c5e6eca114cc27cdd44fc3"} Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.201799 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"3a7cd02a465a6b02f67a7a49f144e54cd8df78eca19e74e08b5718d409010864"} Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.201813 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"f0dc579e5767ea538a9b41ecc841f2689a0f8379a8f13081908698cfa5273fa6"} Dec 06 06:01:33 crc kubenswrapper[4957]: I1206 06:01:33.671829 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pmd7t-config-bq5dj"] Dec 06 06:01:34 crc kubenswrapper[4957]: I1206 06:01:34.211728 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t-config-bq5dj" event={"ID":"06d0255b-0a8d-45f2-acc6-27985b748cad","Type":"ContainerStarted","Data":"d1c2d1a7ea89d67ac386c75749e9ba1345192477f9fcdc631cd84f1472ae331f"} Dec 06 06:01:34 crc kubenswrapper[4957]: I1206 06:01:34.211780 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t-config-bq5dj" event={"ID":"06d0255b-0a8d-45f2-acc6-27985b748cad","Type":"ContainerStarted","Data":"a4e767134f6ba3fd152c39551f03a15a5903ad58d6b4434eeac3f8152174934e"} Dec 06 06:01:34 crc kubenswrapper[4957]: I1206 06:01:34.218181 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"556dd94d7310822297d430410120d683d4946e5d751fa7d6e3ef507a16c2fca3"} Dec 06 06:01:34 crc kubenswrapper[4957]: I1206 06:01:34.233180 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-pmd7t-config-bq5dj" podStartSLOduration=2.233153806 podStartE2EDuration="2.233153806s" podCreationTimestamp="2025-12-06 06:01:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:34.230416822 +0000 UTC m=+1333.880684464" watchObservedRunningTime="2025-12-06 06:01:34.233153806 +0000 UTC m=+1333.883421478" Dec 06 06:01:34 crc kubenswrapper[4957]: I1206 06:01:34.680613 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b64d0f4d-f513-42de-8512-1104a22adb15" path="/var/lib/kubelet/pods/b64d0f4d-f513-42de-8512-1104a22adb15/volumes" Dec 06 06:01:35 crc kubenswrapper[4957]: I1206 06:01:35.243751 4957 generic.go:334] "Generic (PLEG): container finished" podID="06d0255b-0a8d-45f2-acc6-27985b748cad" containerID="d1c2d1a7ea89d67ac386c75749e9ba1345192477f9fcdc631cd84f1472ae331f" exitCode=0 Dec 06 06:01:35 crc kubenswrapper[4957]: I1206 06:01:35.244284 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t-config-bq5dj" event={"ID":"06d0255b-0a8d-45f2-acc6-27985b748cad","Type":"ContainerDied","Data":"d1c2d1a7ea89d67ac386c75749e9ba1345192477f9fcdc631cd84f1472ae331f"} Dec 06 06:01:35 crc kubenswrapper[4957]: I1206 06:01:35.249242 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"029394c34124574bf96e4b59166e1e3cca81a2dedaa14d0b9b7747d2c1dc047e"} Dec 06 06:01:35 crc kubenswrapper[4957]: I1206 06:01:35.249296 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"e7554aefff694f7595a80635b41eb998a19be04e03fbecfb03956ae11901eb6b"} Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.270910 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"abbc45b5dda86d460cc9deb019c67c8323a34bd4c2d7b9e9c7c190845a12d290"} Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.271320 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"55b754b37b1e9a065bbbfd8286c0d6a75f30638f10a67541f52be056f0a44df4"} Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.735089 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.843723 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq755\" (UniqueName: \"kubernetes.io/projected/06d0255b-0a8d-45f2-acc6-27985b748cad-kube-api-access-mq755\") pod \"06d0255b-0a8d-45f2-acc6-27985b748cad\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.843942 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-additional-scripts\") pod \"06d0255b-0a8d-45f2-acc6-27985b748cad\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844075 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-log-ovn\") pod \"06d0255b-0a8d-45f2-acc6-27985b748cad\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844148 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "06d0255b-0a8d-45f2-acc6-27985b748cad" (UID: "06d0255b-0a8d-45f2-acc6-27985b748cad"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844296 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run\") pod \"06d0255b-0a8d-45f2-acc6-27985b748cad\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844389 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run" (OuterVolumeSpecName: "var-run") pod "06d0255b-0a8d-45f2-acc6-27985b748cad" (UID: "06d0255b-0a8d-45f2-acc6-27985b748cad"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844472 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run-ovn\") pod \"06d0255b-0a8d-45f2-acc6-27985b748cad\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844600 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-scripts\") pod \"06d0255b-0a8d-45f2-acc6-27985b748cad\" (UID: \"06d0255b-0a8d-45f2-acc6-27985b748cad\") " Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844497 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "06d0255b-0a8d-45f2-acc6-27985b748cad" (UID: "06d0255b-0a8d-45f2-acc6-27985b748cad"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.844894 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "06d0255b-0a8d-45f2-acc6-27985b748cad" (UID: "06d0255b-0a8d-45f2-acc6-27985b748cad"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.845196 4957 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.845263 4957 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.845328 4957 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.845408 4957 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06d0255b-0a8d-45f2-acc6-27985b748cad-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.845269 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-scripts" (OuterVolumeSpecName: "scripts") pod "06d0255b-0a8d-45f2-acc6-27985b748cad" (UID: "06d0255b-0a8d-45f2-acc6-27985b748cad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.849559 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06d0255b-0a8d-45f2-acc6-27985b748cad-kube-api-access-mq755" (OuterVolumeSpecName: "kube-api-access-mq755") pod "06d0255b-0a8d-45f2-acc6-27985b748cad" (UID: "06d0255b-0a8d-45f2-acc6-27985b748cad"). InnerVolumeSpecName "kube-api-access-mq755". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.946716 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06d0255b-0a8d-45f2-acc6-27985b748cad-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:36 crc kubenswrapper[4957]: I1206 06:01:36.947171 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq755\" (UniqueName: \"kubernetes.io/projected/06d0255b-0a8d-45f2-acc6-27985b748cad-kube-api-access-mq755\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.283083 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pmd7t-config-bq5dj" event={"ID":"06d0255b-0a8d-45f2-acc6-27985b748cad","Type":"ContainerDied","Data":"a4e767134f6ba3fd152c39551f03a15a5903ad58d6b4434eeac3f8152174934e"} Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.283127 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4e767134f6ba3fd152c39551f03a15a5903ad58d6b4434eeac3f8152174934e" Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.283195 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pmd7t-config-bq5dj" Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.288410 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"12a12400a4a76c8211c4df0e7e9876b30834db9aa9e75e59b7c6cbff82b9df86"} Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.288449 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"358401e29626a23bdbedf3206060e9c32fcd85ae26de2cf8686d06ba57c48385"} Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.294738 4957 generic.go:334] "Generic (PLEG): container finished" podID="98dc1a26-e2e0-4447-a841-8507984b8862" containerID="6303e4d02345bcf49a6dfb7dea4862406d20b2007dfb8ddbcd167a45067e42e2" exitCode=0 Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.294789 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scxvl" event={"ID":"98dc1a26-e2e0-4447-a841-8507984b8862","Type":"ContainerDied","Data":"6303e4d02345bcf49a6dfb7dea4862406d20b2007dfb8ddbcd167a45067e42e2"} Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.843196 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-pmd7t-config-bq5dj"] Dec 06 06:01:37 crc kubenswrapper[4957]: I1206 06:01:37.855218 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-pmd7t-config-bq5dj"] Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.148395 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.311955 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"e3e42b2678a02c4d9947f9439c55e7b3d2c331503561b34b8c6a6c49efa72983"} Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.312010 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"1e345ce136d110cdcd012aa4d51ad959c7dc75f4c41e058a8eabceeca1346eba"} Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.312025 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"f3fa15746f4c974da37107f5d919eda5e49c16ea5c78e5ed6e9b44f80cbce0a0"} Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.654690 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.672687 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06d0255b-0a8d-45f2-acc6-27985b748cad" path="/var/lib/kubelet/pods/06d0255b-0a8d-45f2-acc6-27985b748cad/volumes" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.717771 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.786656 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-config-data\") pod \"98dc1a26-e2e0-4447-a841-8507984b8862\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.786707 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62p5s\" (UniqueName: \"kubernetes.io/projected/98dc1a26-e2e0-4447-a841-8507984b8862-kube-api-access-62p5s\") pod \"98dc1a26-e2e0-4447-a841-8507984b8862\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.786777 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-combined-ca-bundle\") pod \"98dc1a26-e2e0-4447-a841-8507984b8862\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.786866 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-db-sync-config-data\") pod \"98dc1a26-e2e0-4447-a841-8507984b8862\" (UID: \"98dc1a26-e2e0-4447-a841-8507984b8862\") " Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.800150 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "98dc1a26-e2e0-4447-a841-8507984b8862" (UID: "98dc1a26-e2e0-4447-a841-8507984b8862"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.800634 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98dc1a26-e2e0-4447-a841-8507984b8862-kube-api-access-62p5s" (OuterVolumeSpecName: "kube-api-access-62p5s") pod "98dc1a26-e2e0-4447-a841-8507984b8862" (UID: "98dc1a26-e2e0-4447-a841-8507984b8862"). InnerVolumeSpecName "kube-api-access-62p5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.828219 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98dc1a26-e2e0-4447-a841-8507984b8862" (UID: "98dc1a26-e2e0-4447-a841-8507984b8862"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.851193 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-config-data" (OuterVolumeSpecName: "config-data") pod "98dc1a26-e2e0-4447-a841-8507984b8862" (UID: "98dc1a26-e2e0-4447-a841-8507984b8862"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.889098 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.889129 4957 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.889139 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98dc1a26-e2e0-4447-a841-8507984b8862-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:38 crc kubenswrapper[4957]: I1206 06:01:38.889148 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62p5s\" (UniqueName: \"kubernetes.io/projected/98dc1a26-e2e0-4447-a841-8507984b8862-kube-api-access-62p5s\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.345578 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"655290396f7d80aa350a26c0a62389c1181248315f0fb4b13e85cfce2c27c3f1"} Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.345816 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"72358664-e14f-482d-bd35-c807922b525f","Type":"ContainerStarted","Data":"3c4b0cf935de8aab2589182e9d3aae6801d596c08de834ce1a554fec884dc573"} Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.351072 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scxvl" event={"ID":"98dc1a26-e2e0-4447-a841-8507984b8862","Type":"ContainerDied","Data":"2b85c5463bf37621af49baf44276a82296892748e42fc854359eb816d08ef976"} Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.351118 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b85c5463bf37621af49baf44276a82296892748e42fc854359eb816d08ef976" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.351210 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scxvl" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.394143 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.023627086 podStartE2EDuration="26.394123654s" podCreationTimestamp="2025-12-06 06:01:13 +0000 UTC" firstStartedPulling="2025-12-06 06:01:31.351078621 +0000 UTC m=+1331.001346253" lastFinishedPulling="2025-12-06 06:01:36.721575179 +0000 UTC m=+1336.371842821" observedRunningTime="2025-12-06 06:01:39.384276327 +0000 UTC m=+1339.034543959" watchObservedRunningTime="2025-12-06 06:01:39.394123654 +0000 UTC m=+1339.044391286" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.707886 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-khpqz"] Dec 06 06:01:39 crc kubenswrapper[4957]: E1206 06:01:39.708285 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98dc1a26-e2e0-4447-a841-8507984b8862" containerName="glance-db-sync" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.708310 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="98dc1a26-e2e0-4447-a841-8507984b8862" containerName="glance-db-sync" Dec 06 06:01:39 crc kubenswrapper[4957]: E1206 06:01:39.708348 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d0255b-0a8d-45f2-acc6-27985b748cad" containerName="ovn-config" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.708357 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d0255b-0a8d-45f2-acc6-27985b748cad" containerName="ovn-config" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.708577 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="98dc1a26-e2e0-4447-a841-8507984b8862" containerName="glance-db-sync" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.708620 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d0255b-0a8d-45f2-acc6-27985b748cad" containerName="ovn-config" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.709744 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.725361 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.727709 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-khpqz"] Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.822497 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.822558 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c5pf\" (UniqueName: \"kubernetes.io/projected/07ab4fa9-6c58-4404-a2b3-479dc532c253-kube-api-access-9c5pf\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.822869 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.822935 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.822956 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-config\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.822977 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-svc\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.854236 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-khpqz"] Dec 06 06:01:39 crc kubenswrapper[4957]: E1206 06:01:39.855005 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-9c5pf ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-764c5664d7-khpqz" podUID="07ab4fa9-6c58-4404-a2b3-479dc532c253" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.913170 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xl9rr"] Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.914702 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.925148 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.925202 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.925224 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-config\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.925244 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-svc\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.925298 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.925329 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c5pf\" (UniqueName: \"kubernetes.io/projected/07ab4fa9-6c58-4404-a2b3-479dc532c253-kube-api-access-9c5pf\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.926710 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.927462 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.928162 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-config\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.928862 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-svc\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.928886 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.972879 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xl9rr"] Dec 06 06:01:39 crc kubenswrapper[4957]: I1206 06:01:39.980951 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c5pf\" (UniqueName: \"kubernetes.io/projected/07ab4fa9-6c58-4404-a2b3-479dc532c253-kube-api-access-9c5pf\") pod \"dnsmasq-dns-764c5664d7-khpqz\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.026760 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.026877 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.026920 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5nrj\" (UniqueName: \"kubernetes.io/projected/06b68237-df64-484c-859a-9dcb92a5059a-kube-api-access-h5nrj\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.027013 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.027044 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-config\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.027109 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.128176 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.128239 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.128256 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5nrj\" (UniqueName: \"kubernetes.io/projected/06b68237-df64-484c-859a-9dcb92a5059a-kube-api-access-h5nrj\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.128313 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.128333 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-config\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.128373 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.129352 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.129941 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.130650 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.130742 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.131288 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-config\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.162662 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5nrj\" (UniqueName: \"kubernetes.io/projected/06b68237-df64-484c-859a-9dcb92a5059a-kube-api-access-h5nrj\") pod \"dnsmasq-dns-74f6bcbc87-xl9rr\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.181126 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.181487 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.262755 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.363434 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.392530 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.437497 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-svc\") pod \"07ab4fa9-6c58-4404-a2b3-479dc532c253\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.437561 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-sb\") pod \"07ab4fa9-6c58-4404-a2b3-479dc532c253\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.437587 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-nb\") pod \"07ab4fa9-6c58-4404-a2b3-479dc532c253\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.437623 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-swift-storage-0\") pod \"07ab4fa9-6c58-4404-a2b3-479dc532c253\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.437681 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c5pf\" (UniqueName: \"kubernetes.io/projected/07ab4fa9-6c58-4404-a2b3-479dc532c253-kube-api-access-9c5pf\") pod \"07ab4fa9-6c58-4404-a2b3-479dc532c253\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.437707 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-config\") pod \"07ab4fa9-6c58-4404-a2b3-479dc532c253\" (UID: \"07ab4fa9-6c58-4404-a2b3-479dc532c253\") " Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.438961 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "07ab4fa9-6c58-4404-a2b3-479dc532c253" (UID: "07ab4fa9-6c58-4404-a2b3-479dc532c253"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.439315 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "07ab4fa9-6c58-4404-a2b3-479dc532c253" (UID: "07ab4fa9-6c58-4404-a2b3-479dc532c253"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.440201 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "07ab4fa9-6c58-4404-a2b3-479dc532c253" (UID: "07ab4fa9-6c58-4404-a2b3-479dc532c253"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.440531 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "07ab4fa9-6c58-4404-a2b3-479dc532c253" (UID: "07ab4fa9-6c58-4404-a2b3-479dc532c253"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.440909 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-config" (OuterVolumeSpecName: "config") pod "07ab4fa9-6c58-4404-a2b3-479dc532c253" (UID: "07ab4fa9-6c58-4404-a2b3-479dc532c253"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.446709 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ab4fa9-6c58-4404-a2b3-479dc532c253-kube-api-access-9c5pf" (OuterVolumeSpecName: "kube-api-access-9c5pf") pod "07ab4fa9-6c58-4404-a2b3-479dc532c253" (UID: "07ab4fa9-6c58-4404-a2b3-479dc532c253"). InnerVolumeSpecName "kube-api-access-9c5pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.540141 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.540180 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.540189 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.540199 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c5pf\" (UniqueName: \"kubernetes.io/projected/07ab4fa9-6c58-4404-a2b3-479dc532c253-kube-api-access-9c5pf\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.540207 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.540216 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07ab4fa9-6c58-4404-a2b3-479dc532c253-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:40 crc kubenswrapper[4957]: I1206 06:01:40.728197 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xl9rr"] Dec 06 06:01:40 crc kubenswrapper[4957]: W1206 06:01:40.731252 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06b68237_df64_484c_859a_9dcb92a5059a.slice/crio-446345d036cbfcb5cabcf1eeecf413531c000e628821d388faf060b37d949146 WatchSource:0}: Error finding container 446345d036cbfcb5cabcf1eeecf413531c000e628821d388faf060b37d949146: Status 404 returned error can't find the container with id 446345d036cbfcb5cabcf1eeecf413531c000e628821d388faf060b37d949146 Dec 06 06:01:41 crc kubenswrapper[4957]: I1206 06:01:41.376633 4957 generic.go:334] "Generic (PLEG): container finished" podID="06b68237-df64-484c-859a-9dcb92a5059a" containerID="8d9885a314fd8362aab5bd731604c4e65aee6a5e03ad4fb94df9824b6245b31f" exitCode=0 Dec 06 06:01:41 crc kubenswrapper[4957]: I1206 06:01:41.376737 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" event={"ID":"06b68237-df64-484c-859a-9dcb92a5059a","Type":"ContainerDied","Data":"8d9885a314fd8362aab5bd731604c4e65aee6a5e03ad4fb94df9824b6245b31f"} Dec 06 06:01:41 crc kubenswrapper[4957]: I1206 06:01:41.376991 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" event={"ID":"06b68237-df64-484c-859a-9dcb92a5059a","Type":"ContainerStarted","Data":"446345d036cbfcb5cabcf1eeecf413531c000e628821d388faf060b37d949146"} Dec 06 06:01:41 crc kubenswrapper[4957]: I1206 06:01:41.377007 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-khpqz" Dec 06 06:01:41 crc kubenswrapper[4957]: I1206 06:01:41.459059 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-khpqz"] Dec 06 06:01:41 crc kubenswrapper[4957]: I1206 06:01:41.479264 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-khpqz"] Dec 06 06:01:42 crc kubenswrapper[4957]: I1206 06:01:42.390104 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" event={"ID":"06b68237-df64-484c-859a-9dcb92a5059a","Type":"ContainerStarted","Data":"94c98f2b707f8620623fbd278850e7ffb3a28d7a1d5786660ad3c897c2fb83e1"} Dec 06 06:01:42 crc kubenswrapper[4957]: I1206 06:01:42.391511 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:42 crc kubenswrapper[4957]: I1206 06:01:42.435660 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" podStartSLOduration=3.43564244 podStartE2EDuration="3.43564244s" podCreationTimestamp="2025-12-06 06:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:42.433246125 +0000 UTC m=+1342.083513757" watchObservedRunningTime="2025-12-06 06:01:42.43564244 +0000 UTC m=+1342.085910072" Dec 06 06:01:42 crc kubenswrapper[4957]: I1206 06:01:42.672535 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07ab4fa9-6c58-4404-a2b3-479dc532c253" path="/var/lib/kubelet/pods/07ab4fa9-6c58-4404-a2b3-479dc532c253/volumes" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.148090 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.567063 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-ptnbn"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.569058 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.577609 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ptnbn"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.655070 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.677697 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-82qnj"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.679567 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.685794 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3e8e-account-create-update-d5w77"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.687091 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.691056 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.702037 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-82qnj"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.713674 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3e8e-account-create-update-d5w77"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.770136 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b916a3e-348a-4b54-8c65-d93e3d684aa8-operator-scripts\") pod \"cinder-db-create-ptnbn\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.770349 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n242m\" (UniqueName: \"kubernetes.io/projected/4b916a3e-348a-4b54-8c65-d93e3d684aa8-kube-api-access-n242m\") pod \"cinder-db-create-ptnbn\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.796249 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-4a8c-account-create-update-hdlfp"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.797237 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.802048 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.854944 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4a8c-account-create-update-hdlfp"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882188 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a963cc41-8008-4761-bf82-84b3e9ab4311-operator-scripts\") pod \"cinder-3e8e-account-create-update-d5w77\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882278 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-operator-scripts\") pod \"barbican-db-create-82qnj\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882302 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pcfx\" (UniqueName: \"kubernetes.io/projected/c491d07c-4774-4670-8467-cd1eed27fd1f-kube-api-access-2pcfx\") pod \"barbican-4a8c-account-create-update-hdlfp\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882343 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b916a3e-348a-4b54-8c65-d93e3d684aa8-operator-scripts\") pod \"cinder-db-create-ptnbn\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882381 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n242m\" (UniqueName: \"kubernetes.io/projected/4b916a3e-348a-4b54-8c65-d93e3d684aa8-kube-api-access-n242m\") pod \"cinder-db-create-ptnbn\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882466 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgljm\" (UniqueName: \"kubernetes.io/projected/a963cc41-8008-4761-bf82-84b3e9ab4311-kube-api-access-mgljm\") pod \"cinder-3e8e-account-create-update-d5w77\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882520 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c491d07c-4774-4670-8467-cd1eed27fd1f-operator-scripts\") pod \"barbican-4a8c-account-create-update-hdlfp\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.882577 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfrr7\" (UniqueName: \"kubernetes.io/projected/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-kube-api-access-lfrr7\") pod \"barbican-db-create-82qnj\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.883440 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b916a3e-348a-4b54-8c65-d93e3d684aa8-operator-scripts\") pod \"cinder-db-create-ptnbn\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.896224 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-zp2bn"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.897455 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.901195 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.901274 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.901336 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hx52c" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.901463 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.904371 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zp2bn"] Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.905397 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n242m\" (UniqueName: \"kubernetes.io/projected/4b916a3e-348a-4b54-8c65-d93e3d684aa8-kube-api-access-n242m\") pod \"cinder-db-create-ptnbn\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984381 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgljm\" (UniqueName: \"kubernetes.io/projected/a963cc41-8008-4761-bf82-84b3e9ab4311-kube-api-access-mgljm\") pod \"cinder-3e8e-account-create-update-d5w77\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984463 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-combined-ca-bundle\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984499 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c491d07c-4774-4670-8467-cd1eed27fd1f-operator-scripts\") pod \"barbican-4a8c-account-create-update-hdlfp\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984551 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfrr7\" (UniqueName: \"kubernetes.io/projected/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-kube-api-access-lfrr7\") pod \"barbican-db-create-82qnj\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984605 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9nhh\" (UniqueName: \"kubernetes.io/projected/7fce1a2e-8245-49f0-a949-091a9165bbc4-kube-api-access-w9nhh\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984642 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a963cc41-8008-4761-bf82-84b3e9ab4311-operator-scripts\") pod \"cinder-3e8e-account-create-update-d5w77\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984689 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-config-data\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984717 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-operator-scripts\") pod \"barbican-db-create-82qnj\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.984740 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pcfx\" (UniqueName: \"kubernetes.io/projected/c491d07c-4774-4670-8467-cd1eed27fd1f-kube-api-access-2pcfx\") pod \"barbican-4a8c-account-create-update-hdlfp\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.985923 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c491d07c-4774-4670-8467-cd1eed27fd1f-operator-scripts\") pod \"barbican-4a8c-account-create-update-hdlfp\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.986694 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a963cc41-8008-4761-bf82-84b3e9ab4311-operator-scripts\") pod \"cinder-3e8e-account-create-update-d5w77\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:48 crc kubenswrapper[4957]: I1206 06:01:48.987501 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-operator-scripts\") pod \"barbican-db-create-82qnj\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.020892 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-m9kbj"] Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.022270 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.027853 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgljm\" (UniqueName: \"kubernetes.io/projected/a963cc41-8008-4761-bf82-84b3e9ab4311-kube-api-access-mgljm\") pod \"cinder-3e8e-account-create-update-d5w77\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.028332 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pcfx\" (UniqueName: \"kubernetes.io/projected/c491d07c-4774-4670-8467-cd1eed27fd1f-kube-api-access-2pcfx\") pod \"barbican-4a8c-account-create-update-hdlfp\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.033747 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfrr7\" (UniqueName: \"kubernetes.io/projected/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-kube-api-access-lfrr7\") pod \"barbican-db-create-82qnj\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.035103 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m9kbj"] Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.088641 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-config-data\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.088716 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8n56\" (UniqueName: \"kubernetes.io/projected/a264853d-5703-430c-ac2d-79d1a1170fa6-kube-api-access-z8n56\") pod \"neutron-db-create-m9kbj\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.088776 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a264853d-5703-430c-ac2d-79d1a1170fa6-operator-scripts\") pod \"neutron-db-create-m9kbj\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.088795 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-combined-ca-bundle\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.088860 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9nhh\" (UniqueName: \"kubernetes.io/projected/7fce1a2e-8245-49f0-a949-091a9165bbc4-kube-api-access-w9nhh\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.095085 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-config-data\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.098702 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8f93-account-create-update-4ktnk"] Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.099055 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-combined-ca-bundle\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.100108 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.104257 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.111237 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8f93-account-create-update-4ktnk"] Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.128913 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9nhh\" (UniqueName: \"kubernetes.io/projected/7fce1a2e-8245-49f0-a949-091a9165bbc4-kube-api-access-w9nhh\") pod \"keystone-db-sync-zp2bn\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.150288 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.191760 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfpl2\" (UniqueName: \"kubernetes.io/projected/9497a720-6416-4d95-87b5-98c6afed7a4c-kube-api-access-kfpl2\") pod \"neutron-8f93-account-create-update-4ktnk\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.191865 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8n56\" (UniqueName: \"kubernetes.io/projected/a264853d-5703-430c-ac2d-79d1a1170fa6-kube-api-access-z8n56\") pod \"neutron-db-create-m9kbj\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.191925 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a264853d-5703-430c-ac2d-79d1a1170fa6-operator-scripts\") pod \"neutron-db-create-m9kbj\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.191955 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9497a720-6416-4d95-87b5-98c6afed7a4c-operator-scripts\") pod \"neutron-8f93-account-create-update-4ktnk\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.192792 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.193163 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a264853d-5703-430c-ac2d-79d1a1170fa6-operator-scripts\") pod \"neutron-db-create-m9kbj\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.215746 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8n56\" (UniqueName: \"kubernetes.io/projected/a264853d-5703-430c-ac2d-79d1a1170fa6-kube-api-access-z8n56\") pod \"neutron-db-create-m9kbj\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.264757 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.296027 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9497a720-6416-4d95-87b5-98c6afed7a4c-operator-scripts\") pod \"neutron-8f93-account-create-update-4ktnk\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.296145 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfpl2\" (UniqueName: \"kubernetes.io/projected/9497a720-6416-4d95-87b5-98c6afed7a4c-kube-api-access-kfpl2\") pod \"neutron-8f93-account-create-update-4ktnk\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.313013 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.315343 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.319872 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9497a720-6416-4d95-87b5-98c6afed7a4c-operator-scripts\") pod \"neutron-8f93-account-create-update-4ktnk\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.354948 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfpl2\" (UniqueName: \"kubernetes.io/projected/9497a720-6416-4d95-87b5-98c6afed7a4c-kube-api-access-kfpl2\") pod \"neutron-8f93-account-create-update-4ktnk\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.512728 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.534326 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.584044 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4a8c-account-create-update-hdlfp"] Dec 06 06:01:49 crc kubenswrapper[4957]: I1206 06:01:49.838051 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ptnbn"] Dec 06 06:01:50 crc kubenswrapper[4957]: I1206 06:01:50.031118 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-82qnj"] Dec 06 06:01:50 crc kubenswrapper[4957]: I1206 06:01:50.035829 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3e8e-account-create-update-d5w77"] Dec 06 06:01:50 crc kubenswrapper[4957]: I1206 06:01:50.042927 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zp2bn"] Dec 06 06:01:50 crc kubenswrapper[4957]: I1206 06:01:50.265078 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:01:50 crc kubenswrapper[4957]: I1206 06:01:50.346053 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5mpr7"] Dec 06 06:01:50 crc kubenswrapper[4957]: I1206 06:01:50.346313 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-5mpr7" podUID="e440d460-8e15-44ea-9500-476aba81d2cf" containerName="dnsmasq-dns" containerID="cri-o://d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075" gracePeriod=10 Dec 06 06:01:52 crc kubenswrapper[4957]: W1206 06:01:52.957996 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc491d07c_4774_4670_8467_cd1eed27fd1f.slice/crio-0a2c1d497224b969a07b05c202d06e8a00435f7239135d712229ac32431a18ca WatchSource:0}: Error finding container 0a2c1d497224b969a07b05c202d06e8a00435f7239135d712229ac32431a18ca: Status 404 returned error can't find the container with id 0a2c1d497224b969a07b05c202d06e8a00435f7239135d712229ac32431a18ca Dec 06 06:01:52 crc kubenswrapper[4957]: W1206 06:01:52.959465 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fce1a2e_8245_49f0_a949_091a9165bbc4.slice/crio-78852fc9b3c4b5f49ded092ce61d7742cb32f2878b84219c5bf52f1c89deb449 WatchSource:0}: Error finding container 78852fc9b3c4b5f49ded092ce61d7742cb32f2878b84219c5bf52f1c89deb449: Status 404 returned error can't find the container with id 78852fc9b3c4b5f49ded092ce61d7742cb32f2878b84219c5bf52f1c89deb449 Dec 06 06:01:52 crc kubenswrapper[4957]: W1206 06:01:52.964014 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda963cc41_8008_4761_bf82_84b3e9ab4311.slice/crio-a877979940369ebd3f76c9f6ca29b25323ec80482d98e27039609d72cfcf1df2 WatchSource:0}: Error finding container a877979940369ebd3f76c9f6ca29b25323ec80482d98e27039609d72cfcf1df2: Status 404 returned error can't find the container with id a877979940369ebd3f76c9f6ca29b25323ec80482d98e27039609d72cfcf1df2 Dec 06 06:01:52 crc kubenswrapper[4957]: W1206 06:01:52.964663 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b916a3e_348a_4b54_8c65_d93e3d684aa8.slice/crio-3188e9640d757052f71f3146e3e7fbdf3dd5782dc716219b2ee54e0a0490d6a8 WatchSource:0}: Error finding container 3188e9640d757052f71f3146e3e7fbdf3dd5782dc716219b2ee54e0a0490d6a8: Status 404 returned error can't find the container with id 3188e9640d757052f71f3146e3e7fbdf3dd5782dc716219b2ee54e0a0490d6a8 Dec 06 06:01:52 crc kubenswrapper[4957]: W1206 06:01:52.974173 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bc78fa5_c7af_4ad7_80aa_2d57b91382f1.slice/crio-1501ea7ba6d45ac25bae1c2f3f4056c6a23c991e31c255ca4935020ac3b14341 WatchSource:0}: Error finding container 1501ea7ba6d45ac25bae1c2f3f4056c6a23c991e31c255ca4935020ac3b14341: Status 404 returned error can't find the container with id 1501ea7ba6d45ac25bae1c2f3f4056c6a23c991e31c255ca4935020ac3b14341 Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.370921 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.418778 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m9kbj"] Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.432186 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8f93-account-create-update-4ktnk"] Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.466493 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-config\") pod \"e440d460-8e15-44ea-9500-476aba81d2cf\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.466543 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-sb\") pod \"e440d460-8e15-44ea-9500-476aba81d2cf\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.466578 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-dns-svc\") pod \"e440d460-8e15-44ea-9500-476aba81d2cf\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.466675 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-nb\") pod \"e440d460-8e15-44ea-9500-476aba81d2cf\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.466720 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnc76\" (UniqueName: \"kubernetes.io/projected/e440d460-8e15-44ea-9500-476aba81d2cf-kube-api-access-cnc76\") pod \"e440d460-8e15-44ea-9500-476aba81d2cf\" (UID: \"e440d460-8e15-44ea-9500-476aba81d2cf\") " Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.473008 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e440d460-8e15-44ea-9500-476aba81d2cf-kube-api-access-cnc76" (OuterVolumeSpecName: "kube-api-access-cnc76") pod "e440d460-8e15-44ea-9500-476aba81d2cf" (UID: "e440d460-8e15-44ea-9500-476aba81d2cf"). InnerVolumeSpecName "kube-api-access-cnc76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.514657 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e440d460-8e15-44ea-9500-476aba81d2cf" (UID: "e440d460-8e15-44ea-9500-476aba81d2cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.517634 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e440d460-8e15-44ea-9500-476aba81d2cf" (UID: "e440d460-8e15-44ea-9500-476aba81d2cf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.522439 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e440d460-8e15-44ea-9500-476aba81d2cf" (UID: "e440d460-8e15-44ea-9500-476aba81d2cf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.527265 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-config" (OuterVolumeSpecName: "config") pod "e440d460-8e15-44ea-9500-476aba81d2cf" (UID: "e440d460-8e15-44ea-9500-476aba81d2cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.539712 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m9kbj" event={"ID":"a264853d-5703-430c-ac2d-79d1a1170fa6","Type":"ContainerStarted","Data":"b0dc4fea87a1862c292cb4462a0c8798c63230538cdd8ebaa044cad05871ee7e"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.541533 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4a8c-account-create-update-hdlfp" event={"ID":"c491d07c-4774-4670-8467-cd1eed27fd1f","Type":"ContainerStarted","Data":"e17712189794c76133c5e4a307894e575a83d7c3d21af732826c132673e8af98"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.541595 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4a8c-account-create-update-hdlfp" event={"ID":"c491d07c-4774-4670-8467-cd1eed27fd1f","Type":"ContainerStarted","Data":"0a2c1d497224b969a07b05c202d06e8a00435f7239135d712229ac32431a18ca"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.543543 4957 generic.go:334] "Generic (PLEG): container finished" podID="e440d460-8e15-44ea-9500-476aba81d2cf" containerID="d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075" exitCode=0 Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.543614 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5mpr7" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.543765 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5mpr7" event={"ID":"e440d460-8e15-44ea-9500-476aba81d2cf","Type":"ContainerDied","Data":"d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.544112 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5mpr7" event={"ID":"e440d460-8e15-44ea-9500-476aba81d2cf","Type":"ContainerDied","Data":"c6b0f032c4b671d7c8c7d3e150dd2bda9bd65e858c554f77158812f4bbc2ac72"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.544140 4957 scope.go:117] "RemoveContainer" containerID="d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.545622 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zp2bn" event={"ID":"7fce1a2e-8245-49f0-a949-091a9165bbc4","Type":"ContainerStarted","Data":"78852fc9b3c4b5f49ded092ce61d7742cb32f2878b84219c5bf52f1c89deb449"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.558139 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-4a8c-account-create-update-hdlfp" podStartSLOduration=5.558119188 podStartE2EDuration="5.558119188s" podCreationTimestamp="2025-12-06 06:01:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:01:53.555457687 +0000 UTC m=+1353.205725319" watchObservedRunningTime="2025-12-06 06:01:53.558119188 +0000 UTC m=+1353.208386820" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.559091 4957 generic.go:334] "Generic (PLEG): container finished" podID="a963cc41-8008-4761-bf82-84b3e9ab4311" containerID="7258bbc3a39b70bee3a5bfb780226cae5c2cee87f1e9957ae4e79ff13c77f180" exitCode=0 Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.559201 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3e8e-account-create-update-d5w77" event={"ID":"a963cc41-8008-4761-bf82-84b3e9ab4311","Type":"ContainerDied","Data":"7258bbc3a39b70bee3a5bfb780226cae5c2cee87f1e9957ae4e79ff13c77f180"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.559235 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3e8e-account-create-update-d5w77" event={"ID":"a963cc41-8008-4761-bf82-84b3e9ab4311","Type":"ContainerStarted","Data":"a877979940369ebd3f76c9f6ca29b25323ec80482d98e27039609d72cfcf1df2"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.563073 4957 generic.go:334] "Generic (PLEG): container finished" podID="7bc78fa5-c7af-4ad7-80aa-2d57b91382f1" containerID="5d8f9614304621184fad18f16681dfdce0daa5a7979ace5e0986af77660eef17" exitCode=0 Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.563149 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-82qnj" event={"ID":"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1","Type":"ContainerDied","Data":"5d8f9614304621184fad18f16681dfdce0daa5a7979ace5e0986af77660eef17"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.563185 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-82qnj" event={"ID":"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1","Type":"ContainerStarted","Data":"1501ea7ba6d45ac25bae1c2f3f4056c6a23c991e31c255ca4935020ac3b14341"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.565169 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8f93-account-create-update-4ktnk" event={"ID":"9497a720-6416-4d95-87b5-98c6afed7a4c","Type":"ContainerStarted","Data":"7c4bc552bc2ed228b70ee379623afb61b70b2e96ca93b9a1f627370896a85b4e"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.567724 4957 generic.go:334] "Generic (PLEG): container finished" podID="4b916a3e-348a-4b54-8c65-d93e3d684aa8" containerID="1007a9cb5ae67115e32f98650fe48e9697435c15a8b3c87316a2bbb566c3ed85" exitCode=0 Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.567764 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ptnbn" event={"ID":"4b916a3e-348a-4b54-8c65-d93e3d684aa8","Type":"ContainerDied","Data":"1007a9cb5ae67115e32f98650fe48e9697435c15a8b3c87316a2bbb566c3ed85"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.567789 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ptnbn" event={"ID":"4b916a3e-348a-4b54-8c65-d93e3d684aa8","Type":"ContainerStarted","Data":"3188e9640d757052f71f3146e3e7fbdf3dd5782dc716219b2ee54e0a0490d6a8"} Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.569473 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.569498 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnc76\" (UniqueName: \"kubernetes.io/projected/e440d460-8e15-44ea-9500-476aba81d2cf-kube-api-access-cnc76\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.569508 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.569518 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.569528 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e440d460-8e15-44ea-9500-476aba81d2cf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.745606 4957 scope.go:117] "RemoveContainer" containerID="bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.783281 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5mpr7"] Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.792647 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5mpr7"] Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.808787 4957 scope.go:117] "RemoveContainer" containerID="d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075" Dec 06 06:01:53 crc kubenswrapper[4957]: E1206 06:01:53.809376 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075\": container with ID starting with d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075 not found: ID does not exist" containerID="d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.809470 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075"} err="failed to get container status \"d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075\": rpc error: code = NotFound desc = could not find container \"d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075\": container with ID starting with d55ad4a25837dd9d2d306ce9e7b3700e95812edc58a7bed6ee977b52c2940075 not found: ID does not exist" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.809512 4957 scope.go:117] "RemoveContainer" containerID="bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9" Dec 06 06:01:53 crc kubenswrapper[4957]: E1206 06:01:53.809962 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9\": container with ID starting with bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9 not found: ID does not exist" containerID="bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9" Dec 06 06:01:53 crc kubenswrapper[4957]: I1206 06:01:53.809993 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9"} err="failed to get container status \"bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9\": rpc error: code = NotFound desc = could not find container \"bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9\": container with ID starting with bd1e292d86aa079584bb0097593e7d7ade725e4e03e7d3ea809a5e803ba0c3e9 not found: ID does not exist" Dec 06 06:01:54 crc kubenswrapper[4957]: I1206 06:01:54.577313 4957 generic.go:334] "Generic (PLEG): container finished" podID="9497a720-6416-4d95-87b5-98c6afed7a4c" containerID="51d293cba8bbe646b4a898c8fb4204635f59dad77bc0cddb72222848af2a81d4" exitCode=0 Dec 06 06:01:54 crc kubenswrapper[4957]: I1206 06:01:54.577375 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8f93-account-create-update-4ktnk" event={"ID":"9497a720-6416-4d95-87b5-98c6afed7a4c","Type":"ContainerDied","Data":"51d293cba8bbe646b4a898c8fb4204635f59dad77bc0cddb72222848af2a81d4"} Dec 06 06:01:54 crc kubenswrapper[4957]: I1206 06:01:54.579945 4957 generic.go:334] "Generic (PLEG): container finished" podID="a264853d-5703-430c-ac2d-79d1a1170fa6" containerID="1eb52ecca5bfcce8f4099786c7777e21f2290a1d3ffec089a7b4e1f40b8ed71c" exitCode=0 Dec 06 06:01:54 crc kubenswrapper[4957]: I1206 06:01:54.580012 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m9kbj" event={"ID":"a264853d-5703-430c-ac2d-79d1a1170fa6","Type":"ContainerDied","Data":"1eb52ecca5bfcce8f4099786c7777e21f2290a1d3ffec089a7b4e1f40b8ed71c"} Dec 06 06:01:54 crc kubenswrapper[4957]: I1206 06:01:54.581883 4957 generic.go:334] "Generic (PLEG): container finished" podID="c491d07c-4774-4670-8467-cd1eed27fd1f" containerID="e17712189794c76133c5e4a307894e575a83d7c3d21af732826c132673e8af98" exitCode=0 Dec 06 06:01:54 crc kubenswrapper[4957]: I1206 06:01:54.581947 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4a8c-account-create-update-hdlfp" event={"ID":"c491d07c-4774-4670-8467-cd1eed27fd1f","Type":"ContainerDied","Data":"e17712189794c76133c5e4a307894e575a83d7c3d21af732826c132673e8af98"} Dec 06 06:01:54 crc kubenswrapper[4957]: I1206 06:01:54.675006 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e440d460-8e15-44ea-9500-476aba81d2cf" path="/var/lib/kubelet/pods/e440d460-8e15-44ea-9500-476aba81d2cf/volumes" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.156610 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.164746 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.175318 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.182554 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.209792 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.212342 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.239589 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b916a3e-348a-4b54-8c65-d93e3d684aa8-operator-scripts\") pod \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.239643 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n242m\" (UniqueName: \"kubernetes.io/projected/4b916a3e-348a-4b54-8c65-d93e3d684aa8-kube-api-access-n242m\") pod \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\" (UID: \"4b916a3e-348a-4b54-8c65-d93e3d684aa8\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.239700 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9497a720-6416-4d95-87b5-98c6afed7a4c-operator-scripts\") pod \"9497a720-6416-4d95-87b5-98c6afed7a4c\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.239748 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfpl2\" (UniqueName: \"kubernetes.io/projected/9497a720-6416-4d95-87b5-98c6afed7a4c-kube-api-access-kfpl2\") pod \"9497a720-6416-4d95-87b5-98c6afed7a4c\" (UID: \"9497a720-6416-4d95-87b5-98c6afed7a4c\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.240542 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b916a3e-348a-4b54-8c65-d93e3d684aa8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4b916a3e-348a-4b54-8c65-d93e3d684aa8" (UID: "4b916a3e-348a-4b54-8c65-d93e3d684aa8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.240994 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9497a720-6416-4d95-87b5-98c6afed7a4c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9497a720-6416-4d95-87b5-98c6afed7a4c" (UID: "9497a720-6416-4d95-87b5-98c6afed7a4c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.246598 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9497a720-6416-4d95-87b5-98c6afed7a4c-kube-api-access-kfpl2" (OuterVolumeSpecName: "kube-api-access-kfpl2") pod "9497a720-6416-4d95-87b5-98c6afed7a4c" (UID: "9497a720-6416-4d95-87b5-98c6afed7a4c"). InnerVolumeSpecName "kube-api-access-kfpl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.261083 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b916a3e-348a-4b54-8c65-d93e3d684aa8-kube-api-access-n242m" (OuterVolumeSpecName: "kube-api-access-n242m") pod "4b916a3e-348a-4b54-8c65-d93e3d684aa8" (UID: "4b916a3e-348a-4b54-8c65-d93e3d684aa8"). InnerVolumeSpecName "kube-api-access-n242m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.341615 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgljm\" (UniqueName: \"kubernetes.io/projected/a963cc41-8008-4761-bf82-84b3e9ab4311-kube-api-access-mgljm\") pod \"a963cc41-8008-4761-bf82-84b3e9ab4311\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.342229 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfrr7\" (UniqueName: \"kubernetes.io/projected/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-kube-api-access-lfrr7\") pod \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.342370 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8n56\" (UniqueName: \"kubernetes.io/projected/a264853d-5703-430c-ac2d-79d1a1170fa6-kube-api-access-z8n56\") pod \"a264853d-5703-430c-ac2d-79d1a1170fa6\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.342508 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pcfx\" (UniqueName: \"kubernetes.io/projected/c491d07c-4774-4670-8467-cd1eed27fd1f-kube-api-access-2pcfx\") pod \"c491d07c-4774-4670-8467-cd1eed27fd1f\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.342710 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a264853d-5703-430c-ac2d-79d1a1170fa6-operator-scripts\") pod \"a264853d-5703-430c-ac2d-79d1a1170fa6\" (UID: \"a264853d-5703-430c-ac2d-79d1a1170fa6\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.342872 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-operator-scripts\") pod \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\" (UID: \"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.343229 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c491d07c-4774-4670-8467-cd1eed27fd1f-operator-scripts\") pod \"c491d07c-4774-4670-8467-cd1eed27fd1f\" (UID: \"c491d07c-4774-4670-8467-cd1eed27fd1f\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.343341 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a264853d-5703-430c-ac2d-79d1a1170fa6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a264853d-5703-430c-ac2d-79d1a1170fa6" (UID: "a264853d-5703-430c-ac2d-79d1a1170fa6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.343535 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7bc78fa5-c7af-4ad7-80aa-2d57b91382f1" (UID: "7bc78fa5-c7af-4ad7-80aa-2d57b91382f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.343811 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c491d07c-4774-4670-8467-cd1eed27fd1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c491d07c-4774-4670-8467-cd1eed27fd1f" (UID: "c491d07c-4774-4670-8467-cd1eed27fd1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.345882 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a963cc41-8008-4761-bf82-84b3e9ab4311-operator-scripts\") pod \"a963cc41-8008-4761-bf82-84b3e9ab4311\" (UID: \"a963cc41-8008-4761-bf82-84b3e9ab4311\") " Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.346328 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a963cc41-8008-4761-bf82-84b3e9ab4311-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a963cc41-8008-4761-bf82-84b3e9ab4311" (UID: "a963cc41-8008-4761-bf82-84b3e9ab4311"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.347199 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a264853d-5703-430c-ac2d-79d1a1170fa6-kube-api-access-z8n56" (OuterVolumeSpecName: "kube-api-access-z8n56") pod "a264853d-5703-430c-ac2d-79d1a1170fa6" (UID: "a264853d-5703-430c-ac2d-79d1a1170fa6"). InnerVolumeSpecName "kube-api-access-z8n56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.348166 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-kube-api-access-lfrr7" (OuterVolumeSpecName: "kube-api-access-lfrr7") pod "7bc78fa5-c7af-4ad7-80aa-2d57b91382f1" (UID: "7bc78fa5-c7af-4ad7-80aa-2d57b91382f1"). InnerVolumeSpecName "kube-api-access-lfrr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.348288 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a963cc41-8008-4761-bf82-84b3e9ab4311-kube-api-access-mgljm" (OuterVolumeSpecName: "kube-api-access-mgljm") pod "a963cc41-8008-4761-bf82-84b3e9ab4311" (UID: "a963cc41-8008-4761-bf82-84b3e9ab4311"). InnerVolumeSpecName "kube-api-access-mgljm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.348306 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c491d07c-4774-4670-8467-cd1eed27fd1f-kube-api-access-2pcfx" (OuterVolumeSpecName: "kube-api-access-2pcfx") pod "c491d07c-4774-4670-8467-cd1eed27fd1f" (UID: "c491d07c-4774-4670-8467-cd1eed27fd1f"). InnerVolumeSpecName "kube-api-access-2pcfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349151 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a963cc41-8008-4761-bf82-84b3e9ab4311-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349182 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgljm\" (UniqueName: \"kubernetes.io/projected/a963cc41-8008-4761-bf82-84b3e9ab4311-kube-api-access-mgljm\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349194 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfrr7\" (UniqueName: \"kubernetes.io/projected/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-kube-api-access-lfrr7\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349206 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8n56\" (UniqueName: \"kubernetes.io/projected/a264853d-5703-430c-ac2d-79d1a1170fa6-kube-api-access-z8n56\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349221 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pcfx\" (UniqueName: \"kubernetes.io/projected/c491d07c-4774-4670-8467-cd1eed27fd1f-kube-api-access-2pcfx\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349233 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a264853d-5703-430c-ac2d-79d1a1170fa6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349245 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349257 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b916a3e-348a-4b54-8c65-d93e3d684aa8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349268 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n242m\" (UniqueName: \"kubernetes.io/projected/4b916a3e-348a-4b54-8c65-d93e3d684aa8-kube-api-access-n242m\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349280 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9497a720-6416-4d95-87b5-98c6afed7a4c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349290 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c491d07c-4774-4670-8467-cd1eed27fd1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.349303 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfpl2\" (UniqueName: \"kubernetes.io/projected/9497a720-6416-4d95-87b5-98c6afed7a4c-kube-api-access-kfpl2\") on node \"crc\" DevicePath \"\"" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.617046 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-82qnj" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.617065 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-82qnj" event={"ID":"7bc78fa5-c7af-4ad7-80aa-2d57b91382f1","Type":"ContainerDied","Data":"1501ea7ba6d45ac25bae1c2f3f4056c6a23c991e31c255ca4935020ac3b14341"} Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.617678 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1501ea7ba6d45ac25bae1c2f3f4056c6a23c991e31c255ca4935020ac3b14341" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.618915 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8f93-account-create-update-4ktnk" event={"ID":"9497a720-6416-4d95-87b5-98c6afed7a4c","Type":"ContainerDied","Data":"7c4bc552bc2ed228b70ee379623afb61b70b2e96ca93b9a1f627370896a85b4e"} Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.619333 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c4bc552bc2ed228b70ee379623afb61b70b2e96ca93b9a1f627370896a85b4e" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.619143 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8f93-account-create-update-4ktnk" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.621071 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ptnbn" event={"ID":"4b916a3e-348a-4b54-8c65-d93e3d684aa8","Type":"ContainerDied","Data":"3188e9640d757052f71f3146e3e7fbdf3dd5782dc716219b2ee54e0a0490d6a8"} Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.621101 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3188e9640d757052f71f3146e3e7fbdf3dd5782dc716219b2ee54e0a0490d6a8" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.621372 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ptnbn" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.622629 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m9kbj" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.622681 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m9kbj" event={"ID":"a264853d-5703-430c-ac2d-79d1a1170fa6","Type":"ContainerDied","Data":"b0dc4fea87a1862c292cb4462a0c8798c63230538cdd8ebaa044cad05871ee7e"} Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.622755 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0dc4fea87a1862c292cb4462a0c8798c63230538cdd8ebaa044cad05871ee7e" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.625822 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4a8c-account-create-update-hdlfp" event={"ID":"c491d07c-4774-4670-8467-cd1eed27fd1f","Type":"ContainerDied","Data":"0a2c1d497224b969a07b05c202d06e8a00435f7239135d712229ac32431a18ca"} Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.625875 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a2c1d497224b969a07b05c202d06e8a00435f7239135d712229ac32431a18ca" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.625884 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4a8c-account-create-update-hdlfp" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.628500 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3e8e-account-create-update-d5w77" event={"ID":"a963cc41-8008-4761-bf82-84b3e9ab4311","Type":"ContainerDied","Data":"a877979940369ebd3f76c9f6ca29b25323ec80482d98e27039609d72cfcf1df2"} Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.628537 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e8e-account-create-update-d5w77" Dec 06 06:01:57 crc kubenswrapper[4957]: I1206 06:01:57.628569 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a877979940369ebd3f76c9f6ca29b25323ec80482d98e27039609d72cfcf1df2" Dec 06 06:01:59 crc kubenswrapper[4957]: I1206 06:01:59.647054 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zp2bn" event={"ID":"7fce1a2e-8245-49f0-a949-091a9165bbc4","Type":"ContainerStarted","Data":"357a6f1005dac5ae3f412795cd9ee52201c5b29114480d41bda62cc119396f97"} Dec 06 06:02:06 crc kubenswrapper[4957]: I1206 06:02:06.732103 4957 generic.go:334] "Generic (PLEG): container finished" podID="7fce1a2e-8245-49f0-a949-091a9165bbc4" containerID="357a6f1005dac5ae3f412795cd9ee52201c5b29114480d41bda62cc119396f97" exitCode=0 Dec 06 06:02:06 crc kubenswrapper[4957]: I1206 06:02:06.732199 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zp2bn" event={"ID":"7fce1a2e-8245-49f0-a949-091a9165bbc4","Type":"ContainerDied","Data":"357a6f1005dac5ae3f412795cd9ee52201c5b29114480d41bda62cc119396f97"} Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.120425 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.238528 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9nhh\" (UniqueName: \"kubernetes.io/projected/7fce1a2e-8245-49f0-a949-091a9165bbc4-kube-api-access-w9nhh\") pod \"7fce1a2e-8245-49f0-a949-091a9165bbc4\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.238719 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-config-data\") pod \"7fce1a2e-8245-49f0-a949-091a9165bbc4\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.238747 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-combined-ca-bundle\") pod \"7fce1a2e-8245-49f0-a949-091a9165bbc4\" (UID: \"7fce1a2e-8245-49f0-a949-091a9165bbc4\") " Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.255114 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce1a2e-8245-49f0-a949-091a9165bbc4-kube-api-access-w9nhh" (OuterVolumeSpecName: "kube-api-access-w9nhh") pod "7fce1a2e-8245-49f0-a949-091a9165bbc4" (UID: "7fce1a2e-8245-49f0-a949-091a9165bbc4"). InnerVolumeSpecName "kube-api-access-w9nhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.282768 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fce1a2e-8245-49f0-a949-091a9165bbc4" (UID: "7fce1a2e-8245-49f0-a949-091a9165bbc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.287882 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-config-data" (OuterVolumeSpecName: "config-data") pod "7fce1a2e-8245-49f0-a949-091a9165bbc4" (UID: "7fce1a2e-8245-49f0-a949-091a9165bbc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.340636 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.340668 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce1a2e-8245-49f0-a949-091a9165bbc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.340697 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9nhh\" (UniqueName: \"kubernetes.io/projected/7fce1a2e-8245-49f0-a949-091a9165bbc4-kube-api-access-w9nhh\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.752184 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zp2bn" event={"ID":"7fce1a2e-8245-49f0-a949-091a9165bbc4","Type":"ContainerDied","Data":"78852fc9b3c4b5f49ded092ce61d7742cb32f2878b84219c5bf52f1c89deb449"} Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.752229 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78852fc9b3c4b5f49ded092ce61d7742cb32f2878b84219c5bf52f1c89deb449" Dec 06 06:02:08 crc kubenswrapper[4957]: I1206 06:02:08.752232 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zp2bn" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.038423 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2g75l"] Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039239 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e440d460-8e15-44ea-9500-476aba81d2cf" containerName="dnsmasq-dns" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039264 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e440d460-8e15-44ea-9500-476aba81d2cf" containerName="dnsmasq-dns" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039511 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b916a3e-348a-4b54-8c65-d93e3d684aa8" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039520 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b916a3e-348a-4b54-8c65-d93e3d684aa8" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039537 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a963cc41-8008-4761-bf82-84b3e9ab4311" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039548 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a963cc41-8008-4761-bf82-84b3e9ab4311" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039565 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fce1a2e-8245-49f0-a949-091a9165bbc4" containerName="keystone-db-sync" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039574 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fce1a2e-8245-49f0-a949-091a9165bbc4" containerName="keystone-db-sync" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039595 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a264853d-5703-430c-ac2d-79d1a1170fa6" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039603 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a264853d-5703-430c-ac2d-79d1a1170fa6" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039623 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9497a720-6416-4d95-87b5-98c6afed7a4c" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039631 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9497a720-6416-4d95-87b5-98c6afed7a4c" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039646 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc78fa5-c7af-4ad7-80aa-2d57b91382f1" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039653 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc78fa5-c7af-4ad7-80aa-2d57b91382f1" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039665 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c491d07c-4774-4670-8467-cd1eed27fd1f" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039672 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c491d07c-4774-4670-8467-cd1eed27fd1f" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: E1206 06:02:09.039685 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e440d460-8e15-44ea-9500-476aba81d2cf" containerName="init" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039692 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e440d460-8e15-44ea-9500-476aba81d2cf" containerName="init" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039913 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c491d07c-4774-4670-8467-cd1eed27fd1f" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039942 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="9497a720-6416-4d95-87b5-98c6afed7a4c" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039953 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e440d460-8e15-44ea-9500-476aba81d2cf" containerName="dnsmasq-dns" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039972 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a963cc41-8008-4761-bf82-84b3e9ab4311" containerName="mariadb-account-create-update" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.039982 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a264853d-5703-430c-ac2d-79d1a1170fa6" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.040005 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fce1a2e-8245-49f0-a949-091a9165bbc4" containerName="keystone-db-sync" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.040017 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc78fa5-c7af-4ad7-80aa-2d57b91382f1" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.040025 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b916a3e-348a-4b54-8c65-d93e3d684aa8" containerName="mariadb-database-create" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.040771 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.043629 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.044192 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.044490 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hx52c" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.045275 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.051699 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.055387 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4ttpg"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.076471 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2g75l"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.076601 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.081203 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4ttpg"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153258 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153328 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-config-data\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153348 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-credential-keys\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153370 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-combined-ca-bundle\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153390 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-scripts\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153419 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153596 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-svc\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153667 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnk8z\" (UniqueName: \"kubernetes.io/projected/178c595d-fae9-4a43-a681-fffbee889dd1-kube-api-access-hnk8z\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153759 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-fernet-keys\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153791 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-config\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153819 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mr9h\" (UniqueName: \"kubernetes.io/projected/ab339550-7a88-4ff8-b421-c9f447624a96-kube-api-access-7mr9h\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.153893 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.207799 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5d74bdd7fc-q6245"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.219534 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.227148 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-h8mhd" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.233383 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.233608 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.233638 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.249615 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d74bdd7fc-q6245"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.255572 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-svc\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.255630 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnk8z\" (UniqueName: \"kubernetes.io/projected/178c595d-fae9-4a43-a681-fffbee889dd1-kube-api-access-hnk8z\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.255673 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-fernet-keys\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.255690 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-config\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.255706 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mr9h\" (UniqueName: \"kubernetes.io/projected/ab339550-7a88-4ff8-b421-c9f447624a96-kube-api-access-7mr9h\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.255724 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256076 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256124 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-config-data\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256142 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-credential-keys\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256161 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-combined-ca-bundle\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256179 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-scripts\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256205 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256696 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-svc\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.256727 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.257177 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-config\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.257377 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.260701 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.266489 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-config-data\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.268537 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-credential-keys\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.269949 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-scripts\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.274700 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-fernet-keys\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.275661 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-combined-ca-bundle\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.285527 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mr9h\" (UniqueName: \"kubernetes.io/projected/ab339550-7a88-4ff8-b421-c9f447624a96-kube-api-access-7mr9h\") pod \"keystone-bootstrap-2g75l\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.293049 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnk8z\" (UniqueName: \"kubernetes.io/projected/178c595d-fae9-4a43-a681-fffbee889dd1-kube-api-access-hnk8z\") pod \"dnsmasq-dns-847c4cc679-4ttpg\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.358933 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c64c7f6-187a-43dc-9994-20efbe94411e-logs\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.358979 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c64c7f6-187a-43dc-9994-20efbe94411e-horizon-secret-key\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.359008 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-config-data\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.359056 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vglc5\" (UniqueName: \"kubernetes.io/projected/8c64c7f6-187a-43dc-9994-20efbe94411e-kube-api-access-vglc5\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.359100 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-scripts\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.359284 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.367636 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-6szxq"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.368631 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.381055 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-h69mr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.381249 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.381364 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.391641 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-gv9kk"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.392605 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.393298 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.395875 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.404224 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.404641 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-srkwj" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.411898 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6szxq"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461288 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-combined-ca-bundle\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461363 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-combined-ca-bundle\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461399 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c64c7f6-187a-43dc-9994-20efbe94411e-logs\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461422 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-config-data\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461450 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c64c7f6-187a-43dc-9994-20efbe94411e-horizon-secret-key\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461474 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-db-sync-config-data\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461500 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-config-data\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461535 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa7934f-ce05-4cac-b198-b67ffa3e96ed-etc-machine-id\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461575 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4qj\" (UniqueName: \"kubernetes.io/projected/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-kube-api-access-qd4qj\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461610 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86chd\" (UniqueName: \"kubernetes.io/projected/caa7934f-ce05-4cac-b198-b67ffa3e96ed-kube-api-access-86chd\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461644 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vglc5\" (UniqueName: \"kubernetes.io/projected/8c64c7f6-187a-43dc-9994-20efbe94411e-kube-api-access-vglc5\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461683 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-config\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461704 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-scripts\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.461727 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-scripts\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.462204 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c64c7f6-187a-43dc-9994-20efbe94411e-logs\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.467071 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-config-data\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.489813 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c64c7f6-187a-43dc-9994-20efbe94411e-horizon-secret-key\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.497028 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-scripts\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.578516 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vglc5\" (UniqueName: \"kubernetes.io/projected/8c64c7f6-187a-43dc-9994-20efbe94411e-kube-api-access-vglc5\") pod \"horizon-5d74bdd7fc-q6245\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584647 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa7934f-ce05-4cac-b198-b67ffa3e96ed-etc-machine-id\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584705 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4qj\" (UniqueName: \"kubernetes.io/projected/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-kube-api-access-qd4qj\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584722 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86chd\" (UniqueName: \"kubernetes.io/projected/caa7934f-ce05-4cac-b198-b67ffa3e96ed-kube-api-access-86chd\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584762 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-config\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584785 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-scripts\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584818 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-combined-ca-bundle\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584864 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-combined-ca-bundle\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584894 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-config-data\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.584917 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-db-sync-config-data\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.586725 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa7934f-ce05-4cac-b198-b67ffa3e96ed-etc-machine-id\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.592186 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-combined-ca-bundle\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.595011 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-db-sync-config-data\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.613668 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-config\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.614361 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-combined-ca-bundle\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.614402 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gv9kk"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.614411 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-config-data\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.615147 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86chd\" (UniqueName: \"kubernetes.io/projected/caa7934f-ce05-4cac-b198-b67ffa3e96ed-kube-api-access-86chd\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.616436 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-scripts\") pod \"cinder-db-sync-6szxq\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.628991 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4qj\" (UniqueName: \"kubernetes.io/projected/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-kube-api-access-qd4qj\") pod \"neutron-db-sync-gv9kk\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.687925 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.715451 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6szxq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.719217 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-f45dr"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.719816 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.731058 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.744772 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.745065 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.745171 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.745509 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d4rs8" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.749200 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.772440 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804349 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hcvj\" (UniqueName: \"kubernetes.io/projected/7d3a3efb-65b9-4a90-bded-99969b537744-kube-api-access-4hcvj\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804462 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-combined-ca-bundle\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804520 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804598 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvxs\" (UniqueName: \"kubernetes.io/projected/3c3e4d9e-461d-468a-a9b2-108124de44ff-kube-api-access-8rvxs\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804630 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-scripts\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804654 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804684 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-run-httpd\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804710 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-log-httpd\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804765 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-db-sync-config-data\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.804811 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-config-data\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.810906 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f45dr"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.820789 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-bcdd755cc-4mppd"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.823930 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.830488 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bcdd755cc-4mppd"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.840482 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.851059 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zsklh"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.853685 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.855965 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.857412 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.857561 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-s85xq" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.866052 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zsklh"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.871285 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4ttpg"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.880066 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qmvb2"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.881555 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.889036 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qmvb2"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.901405 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913391 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-scripts\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913439 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7rsp\" (UniqueName: \"kubernetes.io/projected/430ef178-b9c0-4759-979b-92620ceca29f-kube-api-access-m7rsp\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913458 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-config-data\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913513 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hcvj\" (UniqueName: \"kubernetes.io/projected/7d3a3efb-65b9-4a90-bded-99969b537744-kube-api-access-4hcvj\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913531 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-combined-ca-bundle\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913566 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-combined-ca-bundle\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913583 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-config\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913611 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c167b57e-b623-4080-a770-d87233e00128-logs\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913664 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-config-data\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913717 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913734 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913764 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvxs\" (UniqueName: \"kubernetes.io/projected/3c3e4d9e-461d-468a-a9b2-108124de44ff-kube-api-access-8rvxs\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913796 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-scripts\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913814 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913848 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913876 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-run-httpd\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913895 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-log-httpd\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913924 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430ef178-b9c0-4759-979b-92620ceca29f-logs\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913952 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/430ef178-b9c0-4759-979b-92620ceca29f-horizon-secret-key\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.913986 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-scripts\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.914055 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-db-sync-config-data\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.914136 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv2vv\" (UniqueName: \"kubernetes.io/projected/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-kube-api-access-hv2vv\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.914152 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-config-data\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.914183 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.914205 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6td2m\" (UniqueName: \"kubernetes.io/projected/c167b57e-b623-4080-a770-d87233e00128-kube-api-access-6td2m\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.914286 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.914967 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-run-httpd\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.918242 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-log-httpd\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.922238 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.922235 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.924858 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.925042 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.940670 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-prhgl" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.941527 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.942371 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.946052 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.947451 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hcvj\" (UniqueName: \"kubernetes.io/projected/7d3a3efb-65b9-4a90-bded-99969b537744-kube-api-access-4hcvj\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.948127 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-combined-ca-bundle\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.953175 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-db-sync-config-data\") pod \"barbican-db-sync-f45dr\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.954006 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-scripts\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.954193 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-config-data\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:09 crc kubenswrapper[4957]: I1206 06:02:09.979391 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvxs\" (UniqueName: \"kubernetes.io/projected/3c3e4d9e-461d-468a-a9b2-108124de44ff-kube-api-access-8rvxs\") pod \"ceilometer-0\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " pod="openstack/ceilometer-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015327 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430ef178-b9c0-4759-979b-92620ceca29f-logs\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015376 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/430ef178-b9c0-4759-979b-92620ceca29f-horizon-secret-key\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015400 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-scripts\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015449 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv2vv\" (UniqueName: \"kubernetes.io/projected/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-kube-api-access-hv2vv\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015469 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015487 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6td2m\" (UniqueName: \"kubernetes.io/projected/c167b57e-b623-4080-a770-d87233e00128-kube-api-access-6td2m\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015523 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015547 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-scripts\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015564 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7rsp\" (UniqueName: \"kubernetes.io/projected/430ef178-b9c0-4759-979b-92620ceca29f-kube-api-access-m7rsp\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015583 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-config-data\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015603 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-combined-ca-bundle\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015623 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-config\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015639 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c167b57e-b623-4080-a770-d87233e00128-logs\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015666 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-config-data\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015694 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.015735 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.016543 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.017111 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430ef178-b9c0-4759-979b-92620ceca29f-logs\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.017953 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c167b57e-b623-4080-a770-d87233e00128-logs\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.019797 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.022937 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-scripts\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.024758 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-config-data\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.027179 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-scripts\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.033048 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.034257 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-config\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.034600 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.035979 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-config-data\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.036179 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/430ef178-b9c0-4759-979b-92620ceca29f-horizon-secret-key\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.039699 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-combined-ca-bundle\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.041087 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv2vv\" (UniqueName: \"kubernetes.io/projected/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-kube-api-access-hv2vv\") pod \"dnsmasq-dns-785d8bcb8c-qmvb2\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.044071 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6td2m\" (UniqueName: \"kubernetes.io/projected/c167b57e-b623-4080-a770-d87233e00128-kube-api-access-6td2m\") pod \"placement-db-sync-zsklh\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.044758 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7rsp\" (UniqueName: \"kubernetes.io/projected/430ef178-b9c0-4759-979b-92620ceca29f-kube-api-access-m7rsp\") pod \"horizon-bcdd755cc-4mppd\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.111817 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124040 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124140 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124186 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-scripts\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124217 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-logs\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124245 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qqv5\" (UniqueName: \"kubernetes.io/projected/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-kube-api-access-6qqv5\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124287 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-config-data\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124306 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.124331 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.130252 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f45dr" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.152785 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2g75l"] Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.171516 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.180622 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.180670 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.194092 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.208312 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.229605 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-scripts\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.230219 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-logs\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.230289 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qqv5\" (UniqueName: \"kubernetes.io/projected/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-kube-api-access-6qqv5\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.230804 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-config-data\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.230936 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.235294 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.235637 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.235034 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-logs\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.235995 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.236410 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.236879 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.238802 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.239418 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-config-data\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.240333 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-scripts\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.241413 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:10 crc kubenswrapper[4957]: I1206 06:02:10.252860 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qqv5\" (UniqueName: \"kubernetes.io/projected/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-kube-api-access-6qqv5\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.322061 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.329496 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4ttpg"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.437102 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gv9kk"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.461198 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6szxq"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.502994 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:11 crc kubenswrapper[4957]: W1206 06:02:10.503036 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd90cadb8_4faf_4c1a_b198_0a6b4d537bdc.slice/crio-f891d3bac3b735e01b8c1f10d2a06a52509554b1b9ecdc018113e1d3d7c334a4 WatchSource:0}: Error finding container f891d3bac3b735e01b8c1f10d2a06a52509554b1b9ecdc018113e1d3d7c334a4: Status 404 returned error can't find the container with id f891d3bac3b735e01b8c1f10d2a06a52509554b1b9ecdc018113e1d3d7c334a4 Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.506622 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.510317 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.514644 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.563651 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6pbm\" (UniqueName: \"kubernetes.io/projected/5a0cdab5-b096-405a-9018-9319eccd9b9f-kube-api-access-s6pbm\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.563702 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.563870 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.563952 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.564269 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.564718 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.564804 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.569313 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.573434 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.588432 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: W1206 06:02:10.607743 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c64c7f6_187a_43dc_9994_20efbe94411e.slice/crio-ccfd8a2ee118312161178405c7ccdfbacdbc90cc9e7be6b6a43d76283a35563d WatchSource:0}: Error finding container ccfd8a2ee118312161178405c7ccdfbacdbc90cc9e7be6b6a43d76283a35563d: Status 404 returned error can't find the container with id ccfd8a2ee118312161178405c7ccdfbacdbc90cc9e7be6b6a43d76283a35563d Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.652304 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d74bdd7fc-q6245"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.674860 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.675384 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6pbm\" (UniqueName: \"kubernetes.io/projected/5a0cdab5-b096-405a-9018-9319eccd9b9f-kube-api-access-s6pbm\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.675410 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.675467 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.675483 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.675563 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.675595 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.675628 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.676540 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.680536 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.680563 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.686467 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.687260 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.693594 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.694989 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.749189 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6pbm\" (UniqueName: \"kubernetes.io/projected/5a0cdab5-b096-405a-9018-9319eccd9b9f-kube-api-access-s6pbm\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.758388 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.819003 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2g75l" event={"ID":"ab339550-7a88-4ff8-b421-c9f447624a96","Type":"ContainerStarted","Data":"12f11d69a645f883d92cb5ccc303679687b6eba56bf5fd595ea4fd0d21a1b197"} Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.828609 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gv9kk" event={"ID":"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc","Type":"ContainerStarted","Data":"f891d3bac3b735e01b8c1f10d2a06a52509554b1b9ecdc018113e1d3d7c334a4"} Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.848767 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" event={"ID":"178c595d-fae9-4a43-a681-fffbee889dd1","Type":"ContainerStarted","Data":"7a195ced07da5dae98157fffd29e9f9ab15a92dc3dd3e2878e26a84a75f2e835"} Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.874877 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6szxq" event={"ID":"caa7934f-ce05-4cac-b198-b67ffa3e96ed","Type":"ContainerStarted","Data":"9a443aef8faa446220a003b2dddfb941436fb690ca411191f2a5d71f1df461fc"} Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.899023 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d74bdd7fc-q6245" event={"ID":"8c64c7f6-187a-43dc-9994-20efbe94411e","Type":"ContainerStarted","Data":"ccfd8a2ee118312161178405c7ccdfbacdbc90cc9e7be6b6a43d76283a35563d"} Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:10.971716 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.704989 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.738377 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d74bdd7fc-q6245"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.775802 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.783186 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-54c4596cfc-p2m5g"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.784677 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.823621 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54c4596cfc-p2m5g"] Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.945935 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-scripts\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.946554 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-config-data\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.946668 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90caf120-13de-4d6f-9e95-2954a448e874-logs\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.946728 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw5tn\" (UniqueName: \"kubernetes.io/projected/90caf120-13de-4d6f-9e95-2954a448e874-kube-api-access-dw5tn\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.946813 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90caf120-13de-4d6f-9e95-2954a448e874-horizon-secret-key\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.952641 4957 generic.go:334] "Generic (PLEG): container finished" podID="178c595d-fae9-4a43-a681-fffbee889dd1" containerID="57234fa0e0f23f67c37e7695726a349d8c19e1be26b39b16def2f3c96044eed7" exitCode=0 Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.952775 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" event={"ID":"178c595d-fae9-4a43-a681-fffbee889dd1","Type":"ContainerDied","Data":"57234fa0e0f23f67c37e7695726a349d8c19e1be26b39b16def2f3c96044eed7"} Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.956580 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2g75l" event={"ID":"ab339550-7a88-4ff8-b421-c9f447624a96","Type":"ContainerStarted","Data":"bdf3a62a177b2218ae66053a3efbaaaff377a66d1b101a5263c182572371b465"} Dec 06 06:02:11 crc kubenswrapper[4957]: I1206 06:02:11.968106 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gv9kk" event={"ID":"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc","Type":"ContainerStarted","Data":"956276e152d070d60b34172451f34b8ea8b4798743901cbc7179f04b02bcf206"} Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.021029 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-gv9kk" podStartSLOduration=3.020407085 podStartE2EDuration="3.020407085s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:12.00906071 +0000 UTC m=+1371.659328342" watchObservedRunningTime="2025-12-06 06:02:12.020407085 +0000 UTC m=+1371.670674727" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.035055 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2g75l" podStartSLOduration=3.035035119 podStartE2EDuration="3.035035119s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:12.034123725 +0000 UTC m=+1371.684391377" watchObservedRunningTime="2025-12-06 06:02:12.035035119 +0000 UTC m=+1371.685302771" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.050566 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90caf120-13de-4d6f-9e95-2954a448e874-horizon-secret-key\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.050648 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-scripts\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.050688 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-config-data\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.050779 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90caf120-13de-4d6f-9e95-2954a448e874-logs\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.050819 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw5tn\" (UniqueName: \"kubernetes.io/projected/90caf120-13de-4d6f-9e95-2954a448e874-kube-api-access-dw5tn\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.054745 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-scripts\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.055137 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90caf120-13de-4d6f-9e95-2954a448e874-logs\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.057031 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-config-data\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.069455 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90caf120-13de-4d6f-9e95-2954a448e874-horizon-secret-key\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.088469 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw5tn\" (UniqueName: \"kubernetes.io/projected/90caf120-13de-4d6f-9e95-2954a448e874-kube-api-access-dw5tn\") pod \"horizon-54c4596cfc-p2m5g\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.111221 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.184566 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:02:12 crc kubenswrapper[4957]: W1206 06:02:12.221494 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c3e4d9e_461d_468a_a9b2_108124de44ff.slice/crio-956088ec133a52d47d84d71b74a7bcaf2a43c1cd18b6bef443aac1f95f2889f8 WatchSource:0}: Error finding container 956088ec133a52d47d84d71b74a7bcaf2a43c1cd18b6bef443aac1f95f2889f8: Status 404 returned error can't find the container with id 956088ec133a52d47d84d71b74a7bcaf2a43c1cd18b6bef443aac1f95f2889f8 Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.415263 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.573414 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bcdd755cc-4mppd"] Dec 06 06:02:12 crc kubenswrapper[4957]: W1206 06:02:12.598552 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod430ef178_b9c0_4759_979b_92620ceca29f.slice/crio-cb7aadf5ba55384de2fa9075f7b8fc207c36048ec9832d89a3aff7924021d5ef WatchSource:0}: Error finding container cb7aadf5ba55384de2fa9075f7b8fc207c36048ec9832d89a3aff7924021d5ef: Status 404 returned error can't find the container with id cb7aadf5ba55384de2fa9075f7b8fc207c36048ec9832d89a3aff7924021d5ef Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.647696 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zsklh"] Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.661296 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f45dr"] Dec 06 06:02:12 crc kubenswrapper[4957]: W1206 06:02:12.683615 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc167b57e_b623_4080_a770_d87233e00128.slice/crio-349f856a84776999ffc0f70923b63abe332b236ccaf498cd6177459e9525d863 WatchSource:0}: Error finding container 349f856a84776999ffc0f70923b63abe332b236ccaf498cd6177459e9525d863: Status 404 returned error can't find the container with id 349f856a84776999ffc0f70923b63abe332b236ccaf498cd6177459e9525d863 Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.700336 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qmvb2"] Dec 06 06:02:12 crc kubenswrapper[4957]: W1206 06:02:12.721448 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d3a3efb_65b9_4a90_bded_99969b537744.slice/crio-490412cda4034d2908c83b64415ef0dbffaec5df78c79a63582560494e73612c WatchSource:0}: Error finding container 490412cda4034d2908c83b64415ef0dbffaec5df78c79a63582560494e73612c: Status 404 returned error can't find the container with id 490412cda4034d2908c83b64415ef0dbffaec5df78c79a63582560494e73612c Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.774527 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.783198 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.856708 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54c4596cfc-p2m5g"] Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.877526 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-config\") pod \"178c595d-fae9-4a43-a681-fffbee889dd1\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.877620 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-svc\") pod \"178c595d-fae9-4a43-a681-fffbee889dd1\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.877668 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-sb\") pod \"178c595d-fae9-4a43-a681-fffbee889dd1\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.877697 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-swift-storage-0\") pod \"178c595d-fae9-4a43-a681-fffbee889dd1\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.878432 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-nb\") pod \"178c595d-fae9-4a43-a681-fffbee889dd1\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.879212 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnk8z\" (UniqueName: \"kubernetes.io/projected/178c595d-fae9-4a43-a681-fffbee889dd1-kube-api-access-hnk8z\") pod \"178c595d-fae9-4a43-a681-fffbee889dd1\" (UID: \"178c595d-fae9-4a43-a681-fffbee889dd1\") " Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.893625 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178c595d-fae9-4a43-a681-fffbee889dd1-kube-api-access-hnk8z" (OuterVolumeSpecName: "kube-api-access-hnk8z") pod "178c595d-fae9-4a43-a681-fffbee889dd1" (UID: "178c595d-fae9-4a43-a681-fffbee889dd1"). InnerVolumeSpecName "kube-api-access-hnk8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.901319 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-config" (OuterVolumeSpecName: "config") pod "178c595d-fae9-4a43-a681-fffbee889dd1" (UID: "178c595d-fae9-4a43-a681-fffbee889dd1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.906755 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "178c595d-fae9-4a43-a681-fffbee889dd1" (UID: "178c595d-fae9-4a43-a681-fffbee889dd1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.916143 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "178c595d-fae9-4a43-a681-fffbee889dd1" (UID: "178c595d-fae9-4a43-a681-fffbee889dd1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.927318 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "178c595d-fae9-4a43-a681-fffbee889dd1" (UID: "178c595d-fae9-4a43-a681-fffbee889dd1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.931354 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "178c595d-fae9-4a43-a681-fffbee889dd1" (UID: "178c595d-fae9-4a43-a681-fffbee889dd1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.984081 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.984124 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.984140 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.984154 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.984166 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/178c595d-fae9-4a43-a681-fffbee889dd1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.984177 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnk8z\" (UniqueName: \"kubernetes.io/projected/178c595d-fae9-4a43-a681-fffbee889dd1-kube-api-access-hnk8z\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.987735 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.987741 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-4ttpg" event={"ID":"178c595d-fae9-4a43-a681-fffbee889dd1","Type":"ContainerDied","Data":"7a195ced07da5dae98157fffd29e9f9ab15a92dc3dd3e2878e26a84a75f2e835"} Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.987886 4957 scope.go:117] "RemoveContainer" containerID="57234fa0e0f23f67c37e7695726a349d8c19e1be26b39b16def2f3c96044eed7" Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.991398 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f45dr" event={"ID":"7d3a3efb-65b9-4a90-bded-99969b537744","Type":"ContainerStarted","Data":"490412cda4034d2908c83b64415ef0dbffaec5df78c79a63582560494e73612c"} Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.993674 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerStarted","Data":"956088ec133a52d47d84d71b74a7bcaf2a43c1cd18b6bef443aac1f95f2889f8"} Dec 06 06:02:12 crc kubenswrapper[4957]: I1206 06:02:12.998067 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54c4596cfc-p2m5g" event={"ID":"90caf120-13de-4d6f-9e95-2954a448e874","Type":"ContainerStarted","Data":"3ffaf03690cdea7f1f6824feaad6e3521e68c06166606e8339b02c26e99a5f82"} Dec 06 06:02:13 crc kubenswrapper[4957]: I1206 06:02:13.002204 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" event={"ID":"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e","Type":"ContainerStarted","Data":"1e23d53266c25f9c298b6f716d1ed2d06fb49b0c4528b1876ec27d586c3e5b93"} Dec 06 06:02:13 crc kubenswrapper[4957]: I1206 06:02:13.004849 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd755cc-4mppd" event={"ID":"430ef178-b9c0-4759-979b-92620ceca29f","Type":"ContainerStarted","Data":"cb7aadf5ba55384de2fa9075f7b8fc207c36048ec9832d89a3aff7924021d5ef"} Dec 06 06:02:13 crc kubenswrapper[4957]: I1206 06:02:13.007436 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zsklh" event={"ID":"c167b57e-b623-4080-a770-d87233e00128","Type":"ContainerStarted","Data":"349f856a84776999ffc0f70923b63abe332b236ccaf498cd6177459e9525d863"} Dec 06 06:02:13 crc kubenswrapper[4957]: I1206 06:02:13.011901 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb","Type":"ContainerStarted","Data":"5cd9444ebdcbd7c17bdcd42feb64dd7c825e642b1b48f20fc994dc39cc1b66a6"} Dec 06 06:02:13 crc kubenswrapper[4957]: I1206 06:02:13.064522 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4ttpg"] Dec 06 06:02:13 crc kubenswrapper[4957]: I1206 06:02:13.086547 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4ttpg"] Dec 06 06:02:13 crc kubenswrapper[4957]: I1206 06:02:13.819573 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:13 crc kubenswrapper[4957]: W1206 06:02:13.838204 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a0cdab5_b096_405a_9018_9319eccd9b9f.slice/crio-97752f2c9e232dbbcb18c4553b6d4731971cbb1b74af236badd77745f11eb59c WatchSource:0}: Error finding container 97752f2c9e232dbbcb18c4553b6d4731971cbb1b74af236badd77745f11eb59c: Status 404 returned error can't find the container with id 97752f2c9e232dbbcb18c4553b6d4731971cbb1b74af236badd77745f11eb59c Dec 06 06:02:14 crc kubenswrapper[4957]: I1206 06:02:14.031485 4957 generic.go:334] "Generic (PLEG): container finished" podID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerID="f3866bd21b440a96b64e59c5a687f89e6f231e2bdbec125c33affd83c232ce3d" exitCode=0 Dec 06 06:02:14 crc kubenswrapper[4957]: I1206 06:02:14.031719 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" event={"ID":"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e","Type":"ContainerDied","Data":"f3866bd21b440a96b64e59c5a687f89e6f231e2bdbec125c33affd83c232ce3d"} Dec 06 06:02:14 crc kubenswrapper[4957]: I1206 06:02:14.039664 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb","Type":"ContainerStarted","Data":"33b22b04c08c8a56fe1ebe806b229b124fcebfc0fe4f405e0c3ae983be758002"} Dec 06 06:02:14 crc kubenswrapper[4957]: I1206 06:02:14.045391 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a0cdab5-b096-405a-9018-9319eccd9b9f","Type":"ContainerStarted","Data":"97752f2c9e232dbbcb18c4553b6d4731971cbb1b74af236badd77745f11eb59c"} Dec 06 06:02:14 crc kubenswrapper[4957]: I1206 06:02:14.683238 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="178c595d-fae9-4a43-a681-fffbee889dd1" path="/var/lib/kubelet/pods/178c595d-fae9-4a43-a681-fffbee889dd1/volumes" Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.059010 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a0cdab5-b096-405a-9018-9319eccd9b9f","Type":"ContainerStarted","Data":"92947cccf0669f5e9c7584d557033334706aa9b8d23cc7787538bea6adb644fa"} Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.065660 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" event={"ID":"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e","Type":"ContainerStarted","Data":"0d35b8b3da1aab99f8dc72cf460458a1afe19bd0cdc7862664f55a885a4bbc56"} Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.066163 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.074417 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb","Type":"ContainerStarted","Data":"7a97b76267c3521c8ee187c20e74040fc6329c61fd79598e0a5784b5d5bfdab6"} Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.074622 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-log" containerID="cri-o://33b22b04c08c8a56fe1ebe806b229b124fcebfc0fe4f405e0c3ae983be758002" gracePeriod=30 Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.074766 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-httpd" containerID="cri-o://7a97b76267c3521c8ee187c20e74040fc6329c61fd79598e0a5784b5d5bfdab6" gracePeriod=30 Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.097261 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" podStartSLOduration=6.097231506 podStartE2EDuration="6.097231506s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:15.088301906 +0000 UTC m=+1374.738569558" watchObservedRunningTime="2025-12-06 06:02:15.097231506 +0000 UTC m=+1374.747499168" Dec 06 06:02:15 crc kubenswrapper[4957]: I1206 06:02:15.124454 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.124428539 podStartE2EDuration="6.124428539s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:15.115367875 +0000 UTC m=+1374.765635527" watchObservedRunningTime="2025-12-06 06:02:15.124428539 +0000 UTC m=+1374.774696171" Dec 06 06:02:16 crc kubenswrapper[4957]: I1206 06:02:16.087731 4957 generic.go:334] "Generic (PLEG): container finished" podID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerID="7a97b76267c3521c8ee187c20e74040fc6329c61fd79598e0a5784b5d5bfdab6" exitCode=0 Dec 06 06:02:16 crc kubenswrapper[4957]: I1206 06:02:16.088156 4957 generic.go:334] "Generic (PLEG): container finished" podID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerID="33b22b04c08c8a56fe1ebe806b229b124fcebfc0fe4f405e0c3ae983be758002" exitCode=143 Dec 06 06:02:16 crc kubenswrapper[4957]: I1206 06:02:16.087912 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb","Type":"ContainerDied","Data":"7a97b76267c3521c8ee187c20e74040fc6329c61fd79598e0a5784b5d5bfdab6"} Dec 06 06:02:16 crc kubenswrapper[4957]: I1206 06:02:16.089134 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb","Type":"ContainerDied","Data":"33b22b04c08c8a56fe1ebe806b229b124fcebfc0fe4f405e0c3ae983be758002"} Dec 06 06:02:17 crc kubenswrapper[4957]: I1206 06:02:17.100787 4957 generic.go:334] "Generic (PLEG): container finished" podID="ab339550-7a88-4ff8-b421-c9f447624a96" containerID="bdf3a62a177b2218ae66053a3efbaaaff377a66d1b101a5263c182572371b465" exitCode=0 Dec 06 06:02:17 crc kubenswrapper[4957]: I1206 06:02:17.100875 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2g75l" event={"ID":"ab339550-7a88-4ff8-b421-c9f447624a96","Type":"ContainerDied","Data":"bdf3a62a177b2218ae66053a3efbaaaff377a66d1b101a5263c182572371b465"} Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.408444 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bcdd755cc-4mppd"] Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.456351 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-76c6cdbd84-nmkc9"] Dec 06 06:02:18 crc kubenswrapper[4957]: E1206 06:02:18.457153 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178c595d-fae9-4a43-a681-fffbee889dd1" containerName="init" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.457167 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="178c595d-fae9-4a43-a681-fffbee889dd1" containerName="init" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.457563 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="178c595d-fae9-4a43-a681-fffbee889dd1" containerName="init" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.459604 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.464928 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493119 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-secret-key\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493197 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-tls-certs\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493237 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-combined-ca-bundle\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493279 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-scripts\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493295 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76c6cdbd84-nmkc9"] Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493306 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d4332b6-8252-435b-9a1c-bc3f27addaea-logs\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493388 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-config-data\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.493426 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hjkt\" (UniqueName: \"kubernetes.io/projected/3d4332b6-8252-435b-9a1c-bc3f27addaea-kube-api-access-9hjkt\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.595003 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-secret-key\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.595319 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-tls-certs\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.595422 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-combined-ca-bundle\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.595532 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-scripts\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.595618 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d4332b6-8252-435b-9a1c-bc3f27addaea-logs\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.595705 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-config-data\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.595808 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hjkt\" (UniqueName: \"kubernetes.io/projected/3d4332b6-8252-435b-9a1c-bc3f27addaea-kube-api-access-9hjkt\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.598599 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-scripts\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.599220 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d4332b6-8252-435b-9a1c-bc3f27addaea-logs\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.600384 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-config-data\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.605753 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-combined-ca-bundle\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.609378 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-tls-certs\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.620746 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-secret-key\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.624594 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54c4596cfc-p2m5g"] Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.629911 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hjkt\" (UniqueName: \"kubernetes.io/projected/3d4332b6-8252-435b-9a1c-bc3f27addaea-kube-api-access-9hjkt\") pod \"horizon-76c6cdbd84-nmkc9\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.717011 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c76d956-bsr6l"] Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.718991 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.765268 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c76d956-bsr6l"] Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.787334 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.908409 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-horizon-tls-certs\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.908507 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb06a82d-1610-4985-8168-c95f32ad2ff6-config-data\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.908530 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjrfm\" (UniqueName: \"kubernetes.io/projected/cb06a82d-1610-4985-8168-c95f32ad2ff6-kube-api-access-qjrfm\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.908604 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-combined-ca-bundle\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.908671 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-horizon-secret-key\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.908699 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb06a82d-1610-4985-8168-c95f32ad2ff6-logs\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:18 crc kubenswrapper[4957]: I1206 06:02:18.908866 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb06a82d-1610-4985-8168-c95f32ad2ff6-scripts\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.011206 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb06a82d-1610-4985-8168-c95f32ad2ff6-scripts\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.011398 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-horizon-tls-certs\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.011431 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb06a82d-1610-4985-8168-c95f32ad2ff6-config-data\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.011456 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjrfm\" (UniqueName: \"kubernetes.io/projected/cb06a82d-1610-4985-8168-c95f32ad2ff6-kube-api-access-qjrfm\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.011482 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-combined-ca-bundle\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.011522 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-horizon-secret-key\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.011563 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb06a82d-1610-4985-8168-c95f32ad2ff6-logs\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.012569 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb06a82d-1610-4985-8168-c95f32ad2ff6-scripts\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.012765 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb06a82d-1610-4985-8168-c95f32ad2ff6-logs\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.016510 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb06a82d-1610-4985-8168-c95f32ad2ff6-config-data\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.018797 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-horizon-secret-key\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.018901 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-horizon-tls-certs\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.019526 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb06a82d-1610-4985-8168-c95f32ad2ff6-combined-ca-bundle\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.029096 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjrfm\" (UniqueName: \"kubernetes.io/projected/cb06a82d-1610-4985-8168-c95f32ad2ff6-kube-api-access-qjrfm\") pod \"horizon-5c76d956-bsr6l\" (UID: \"cb06a82d-1610-4985-8168-c95f32ad2ff6\") " pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:19 crc kubenswrapper[4957]: I1206 06:02:19.090251 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:20 crc kubenswrapper[4957]: I1206 06:02:20.210685 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:20 crc kubenswrapper[4957]: I1206 06:02:20.313724 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xl9rr"] Dec 06 06:02:20 crc kubenswrapper[4957]: I1206 06:02:20.313962 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="dnsmasq-dns" containerID="cri-o://94c98f2b707f8620623fbd278850e7ffb3a28d7a1d5786660ad3c897c2fb83e1" gracePeriod=10 Dec 06 06:02:21 crc kubenswrapper[4957]: I1206 06:02:21.151612 4957 generic.go:334] "Generic (PLEG): container finished" podID="06b68237-df64-484c-859a-9dcb92a5059a" containerID="94c98f2b707f8620623fbd278850e7ffb3a28d7a1d5786660ad3c897c2fb83e1" exitCode=0 Dec 06 06:02:21 crc kubenswrapper[4957]: I1206 06:02:21.151928 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" event={"ID":"06b68237-df64-484c-859a-9dcb92a5059a","Type":"ContainerDied","Data":"94c98f2b707f8620623fbd278850e7ffb3a28d7a1d5786660ad3c897c2fb83e1"} Dec 06 06:02:30 crc kubenswrapper[4957]: I1206 06:02:30.264009 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.265444 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 06 06:02:35 crc kubenswrapper[4957]: E1206 06:02:35.544529 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 06 06:02:35 crc kubenswrapper[4957]: E1206 06:02:35.545014 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6td2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-zsklh_openstack(c167b57e-b623-4080-a770-d87233e00128): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:02:35 crc kubenswrapper[4957]: E1206 06:02:35.546588 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-zsklh" podUID="c167b57e-b623-4080-a770-d87233e00128" Dec 06 06:02:35 crc kubenswrapper[4957]: E1206 06:02:35.561282 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 06 06:02:35 crc kubenswrapper[4957]: E1206 06:02:35.561515 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd8h59h564h94h5f7h655h9h5f9h688h88hfch7h655h8fh5fch5f8hf8h54h568h57h549hb4h66chbbh75h58dh67hch676h666h68ch645q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vglc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5d74bdd7fc-q6245_openstack(8c64c7f6-187a-43dc-9994-20efbe94411e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:02:35 crc kubenswrapper[4957]: E1206 06:02:35.564262 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5d74bdd7fc-q6245" podUID="8c64c7f6-187a-43dc-9994-20efbe94411e" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.636275 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.642109 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831003 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-combined-ca-bundle\") pod \"ab339550-7a88-4ff8-b421-c9f447624a96\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831054 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-credential-keys\") pod \"ab339550-7a88-4ff8-b421-c9f447624a96\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831080 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-public-tls-certs\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831105 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831169 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-combined-ca-bundle\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831191 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-scripts\") pod \"ab339550-7a88-4ff8-b421-c9f447624a96\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831236 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-config-data\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831262 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-config-data\") pod \"ab339550-7a88-4ff8-b421-c9f447624a96\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831279 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qqv5\" (UniqueName: \"kubernetes.io/projected/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-kube-api-access-6qqv5\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831323 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-logs\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831353 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-httpd-run\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831377 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-fernet-keys\") pod \"ab339550-7a88-4ff8-b421-c9f447624a96\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831396 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mr9h\" (UniqueName: \"kubernetes.io/projected/ab339550-7a88-4ff8-b421-c9f447624a96-kube-api-access-7mr9h\") pod \"ab339550-7a88-4ff8-b421-c9f447624a96\" (UID: \"ab339550-7a88-4ff8-b421-c9f447624a96\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.831413 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-scripts\") pod \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\" (UID: \"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb\") " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.833764 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.834747 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-logs" (OuterVolumeSpecName: "logs") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.838030 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ab339550-7a88-4ff8-b421-c9f447624a96" (UID: "ab339550-7a88-4ff8-b421-c9f447624a96"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.838210 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ab339550-7a88-4ff8-b421-c9f447624a96" (UID: "ab339550-7a88-4ff8-b421-c9f447624a96"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.838276 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-kube-api-access-6qqv5" (OuterVolumeSpecName: "kube-api-access-6qqv5") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "kube-api-access-6qqv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.841464 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-scripts" (OuterVolumeSpecName: "scripts") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.841590 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-scripts" (OuterVolumeSpecName: "scripts") pod "ab339550-7a88-4ff8-b421-c9f447624a96" (UID: "ab339550-7a88-4ff8-b421-c9f447624a96"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.842096 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.843856 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab339550-7a88-4ff8-b421-c9f447624a96-kube-api-access-7mr9h" (OuterVolumeSpecName: "kube-api-access-7mr9h") pod "ab339550-7a88-4ff8-b421-c9f447624a96" (UID: "ab339550-7a88-4ff8-b421-c9f447624a96"). InnerVolumeSpecName "kube-api-access-7mr9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.864250 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.867178 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab339550-7a88-4ff8-b421-c9f447624a96" (UID: "ab339550-7a88-4ff8-b421-c9f447624a96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.871504 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-config-data" (OuterVolumeSpecName: "config-data") pod "ab339550-7a88-4ff8-b421-c9f447624a96" (UID: "ab339550-7a88-4ff8-b421-c9f447624a96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.885271 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.889967 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-config-data" (OuterVolumeSpecName: "config-data") pod "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" (UID: "cdc0b95f-5b60-4b36-a5e3-244b892d8bbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933530 4957 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933577 4957 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933621 4957 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933633 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933645 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933656 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933667 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933677 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qqv5\" (UniqueName: \"kubernetes.io/projected/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-kube-api-access-6qqv5\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933690 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933701 4957 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933711 4957 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933721 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mr9h\" (UniqueName: \"kubernetes.io/projected/ab339550-7a88-4ff8-b421-c9f447624a96-kube-api-access-7mr9h\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933733 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.933744 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab339550-7a88-4ff8-b421-c9f447624a96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:35 crc kubenswrapper[4957]: I1206 06:02:35.951696 4957 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.035583 4957 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.295272 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cdc0b95f-5b60-4b36-a5e3-244b892d8bbb","Type":"ContainerDied","Data":"5cd9444ebdcbd7c17bdcd42feb64dd7c825e642b1b48f20fc994dc39cc1b66a6"} Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.295326 4957 scope.go:117] "RemoveContainer" containerID="7a97b76267c3521c8ee187c20e74040fc6329c61fd79598e0a5784b5d5bfdab6" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.295651 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.298092 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2g75l" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.298821 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2g75l" event={"ID":"ab339550-7a88-4ff8-b421-c9f447624a96","Type":"ContainerDied","Data":"12f11d69a645f883d92cb5ccc303679687b6eba56bf5fd595ea4fd0d21a1b197"} Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.298870 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12f11d69a645f883d92cb5ccc303679687b6eba56bf5fd595ea4fd0d21a1b197" Dec 06 06:02:36 crc kubenswrapper[4957]: E1206 06:02:36.302567 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-zsklh" podUID="c167b57e-b623-4080-a770-d87233e00128" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.392537 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.405019 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.416180 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:36 crc kubenswrapper[4957]: E1206 06:02:36.416533 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-log" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.416549 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-log" Dec 06 06:02:36 crc kubenswrapper[4957]: E1206 06:02:36.416565 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab339550-7a88-4ff8-b421-c9f447624a96" containerName="keystone-bootstrap" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.416571 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab339550-7a88-4ff8-b421-c9f447624a96" containerName="keystone-bootstrap" Dec 06 06:02:36 crc kubenswrapper[4957]: E1206 06:02:36.416587 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-httpd" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.416593 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-httpd" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.416770 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-log" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.416793 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab339550-7a88-4ff8-b421-c9f447624a96" containerName="keystone-bootstrap" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.416806 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" containerName="glance-httpd" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.417800 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.419954 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.420921 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.424784 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548394 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-logs\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548460 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548490 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-scripts\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548572 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548673 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548712 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548761 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4thw\" (UniqueName: \"kubernetes.io/projected/376b5f5c-4e23-4e75-9619-698a4d95eab5-kube-api-access-n4thw\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.548824 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-config-data\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650387 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-logs\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650443 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650467 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-scripts\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650496 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650579 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650615 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650657 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4thw\" (UniqueName: \"kubernetes.io/projected/376b5f5c-4e23-4e75-9619-698a4d95eab5-kube-api-access-n4thw\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.650710 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-config-data\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.651572 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-logs\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.651713 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.651781 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.656191 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.656676 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.667615 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-scripts\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.667794 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4thw\" (UniqueName: \"kubernetes.io/projected/376b5f5c-4e23-4e75-9619-698a4d95eab5-kube-api-access-n4thw\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.673098 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-config-data\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.676626 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc0b95f-5b60-4b36-a5e3-244b892d8bbb" path="/var/lib/kubelet/pods/cdc0b95f-5b60-4b36-a5e3-244b892d8bbb/volumes" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.679141 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.740432 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2g75l"] Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.749739 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2g75l"] Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.752970 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.835664 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bqh2v"] Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.837244 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.841818 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.842077 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.842149 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hx52c" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.842237 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.842401 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.846211 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bqh2v"] Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.956037 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-fernet-keys\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.956191 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-combined-ca-bundle\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.956516 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-scripts\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.956579 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-credential-keys\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.956631 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-config-data\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:36 crc kubenswrapper[4957]: I1206 06:02:36.956732 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq9dn\" (UniqueName: \"kubernetes.io/projected/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-kube-api-access-pq9dn\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.058442 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-combined-ca-bundle\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.058540 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-scripts\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.058565 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-credential-keys\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.058587 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-config-data\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.058622 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq9dn\" (UniqueName: \"kubernetes.io/projected/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-kube-api-access-pq9dn\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.058699 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-fernet-keys\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.063657 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-config-data\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.064948 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-credential-keys\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.065112 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-combined-ca-bundle\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.066176 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-scripts\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.075666 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq9dn\" (UniqueName: \"kubernetes.io/projected/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-kube-api-access-pq9dn\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.077219 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-fernet-keys\") pod \"keystone-bootstrap-bqh2v\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.154717 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.307413 4957 generic.go:334] "Generic (PLEG): container finished" podID="d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" containerID="956276e152d070d60b34172451f34b8ea8b4798743901cbc7179f04b02bcf206" exitCode=0 Dec 06 06:02:37 crc kubenswrapper[4957]: I1206 06:02:37.307455 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gv9kk" event={"ID":"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc","Type":"ContainerDied","Data":"956276e152d070d60b34172451f34b8ea8b4798743901cbc7179f04b02bcf206"} Dec 06 06:02:38 crc kubenswrapper[4957]: E1206 06:02:38.145560 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 06 06:02:38 crc kubenswrapper[4957]: E1206 06:02:38.145905 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64dh674h595h655h55ch7h649h5bh5c6hcdh5bh644h9fh87h59dh646h58dh577hb6h669h655h586h574h5b6h56bh5cfh5c5h65hcdh585h695h65q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8rvxs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(3c3e4d9e-461d-468a-a9b2-108124de44ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.153413 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.233377 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.316571 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.316582 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" event={"ID":"06b68237-df64-484c-859a-9dcb92a5059a","Type":"ContainerDied","Data":"446345d036cbfcb5cabcf1eeecf413531c000e628821d388faf060b37d949146"} Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.384416 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-sb\") pod \"06b68237-df64-484c-859a-9dcb92a5059a\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.384604 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-config\") pod \"06b68237-df64-484c-859a-9dcb92a5059a\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.384644 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-svc\") pod \"06b68237-df64-484c-859a-9dcb92a5059a\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.384693 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-swift-storage-0\") pod \"06b68237-df64-484c-859a-9dcb92a5059a\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.384765 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5nrj\" (UniqueName: \"kubernetes.io/projected/06b68237-df64-484c-859a-9dcb92a5059a-kube-api-access-h5nrj\") pod \"06b68237-df64-484c-859a-9dcb92a5059a\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.384916 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-nb\") pod \"06b68237-df64-484c-859a-9dcb92a5059a\" (UID: \"06b68237-df64-484c-859a-9dcb92a5059a\") " Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.399228 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06b68237-df64-484c-859a-9dcb92a5059a-kube-api-access-h5nrj" (OuterVolumeSpecName: "kube-api-access-h5nrj") pod "06b68237-df64-484c-859a-9dcb92a5059a" (UID: "06b68237-df64-484c-859a-9dcb92a5059a"). InnerVolumeSpecName "kube-api-access-h5nrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.440641 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "06b68237-df64-484c-859a-9dcb92a5059a" (UID: "06b68237-df64-484c-859a-9dcb92a5059a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.444387 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "06b68237-df64-484c-859a-9dcb92a5059a" (UID: "06b68237-df64-484c-859a-9dcb92a5059a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.455672 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "06b68237-df64-484c-859a-9dcb92a5059a" (UID: "06b68237-df64-484c-859a-9dcb92a5059a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.456699 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-config" (OuterVolumeSpecName: "config") pod "06b68237-df64-484c-859a-9dcb92a5059a" (UID: "06b68237-df64-484c-859a-9dcb92a5059a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.467430 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "06b68237-df64-484c-859a-9dcb92a5059a" (UID: "06b68237-df64-484c-859a-9dcb92a5059a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.487945 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5nrj\" (UniqueName: \"kubernetes.io/projected/06b68237-df64-484c-859a-9dcb92a5059a-kube-api-access-h5nrj\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.487983 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.487992 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.488001 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.488010 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.488018 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06b68237-df64-484c-859a-9dcb92a5059a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.653024 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xl9rr"] Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.660129 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-xl9rr"] Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.672552 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06b68237-df64-484c-859a-9dcb92a5059a" path="/var/lib/kubelet/pods/06b68237-df64-484c-859a-9dcb92a5059a/volumes" Dec 06 06:02:38 crc kubenswrapper[4957]: I1206 06:02:38.673806 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab339550-7a88-4ff8-b421-c9f447624a96" path="/var/lib/kubelet/pods/ab339550-7a88-4ff8-b421-c9f447624a96/volumes" Dec 06 06:02:39 crc kubenswrapper[4957]: E1206 06:02:39.233349 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 06 06:02:39 crc kubenswrapper[4957]: E1206 06:02:39.233545 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86chd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-6szxq_openstack(caa7934f-ce05-4cac-b198-b67ffa3e96ed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:02:39 crc kubenswrapper[4957]: E1206 06:02:39.234810 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-6szxq" podUID="caa7934f-ce05-4cac-b198-b67ffa3e96ed" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.258268 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.325220 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d74bdd7fc-q6245" event={"ID":"8c64c7f6-187a-43dc-9994-20efbe94411e","Type":"ContainerDied","Data":"ccfd8a2ee118312161178405c7ccdfbacdbc90cc9e7be6b6a43d76283a35563d"} Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.325406 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d74bdd7fc-q6245" Dec 06 06:02:39 crc kubenswrapper[4957]: E1206 06:02:39.326623 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-6szxq" podUID="caa7934f-ce05-4cac-b198-b67ffa3e96ed" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.410459 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c64c7f6-187a-43dc-9994-20efbe94411e-logs\") pod \"8c64c7f6-187a-43dc-9994-20efbe94411e\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.410509 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-config-data\") pod \"8c64c7f6-187a-43dc-9994-20efbe94411e\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.410558 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c64c7f6-187a-43dc-9994-20efbe94411e-horizon-secret-key\") pod \"8c64c7f6-187a-43dc-9994-20efbe94411e\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.410678 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-scripts\") pod \"8c64c7f6-187a-43dc-9994-20efbe94411e\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.410731 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vglc5\" (UniqueName: \"kubernetes.io/projected/8c64c7f6-187a-43dc-9994-20efbe94411e-kube-api-access-vglc5\") pod \"8c64c7f6-187a-43dc-9994-20efbe94411e\" (UID: \"8c64c7f6-187a-43dc-9994-20efbe94411e\") " Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.410788 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c64c7f6-187a-43dc-9994-20efbe94411e-logs" (OuterVolumeSpecName: "logs") pod "8c64c7f6-187a-43dc-9994-20efbe94411e" (UID: "8c64c7f6-187a-43dc-9994-20efbe94411e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.411223 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c64c7f6-187a-43dc-9994-20efbe94411e-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.411423 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-config-data" (OuterVolumeSpecName: "config-data") pod "8c64c7f6-187a-43dc-9994-20efbe94411e" (UID: "8c64c7f6-187a-43dc-9994-20efbe94411e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.411603 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-scripts" (OuterVolumeSpecName: "scripts") pod "8c64c7f6-187a-43dc-9994-20efbe94411e" (UID: "8c64c7f6-187a-43dc-9994-20efbe94411e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.414902 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c64c7f6-187a-43dc-9994-20efbe94411e-kube-api-access-vglc5" (OuterVolumeSpecName: "kube-api-access-vglc5") pod "8c64c7f6-187a-43dc-9994-20efbe94411e" (UID: "8c64c7f6-187a-43dc-9994-20efbe94411e"). InnerVolumeSpecName "kube-api-access-vglc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.415773 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c64c7f6-187a-43dc-9994-20efbe94411e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8c64c7f6-187a-43dc-9994-20efbe94411e" (UID: "8c64c7f6-187a-43dc-9994-20efbe94411e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.514266 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.514307 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vglc5\" (UniqueName: \"kubernetes.io/projected/8c64c7f6-187a-43dc-9994-20efbe94411e-kube-api-access-vglc5\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.514321 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c64c7f6-187a-43dc-9994-20efbe94411e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.514333 4957 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c64c7f6-187a-43dc-9994-20efbe94411e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.678066 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d74bdd7fc-q6245"] Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.684732 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5d74bdd7fc-q6245"] Dec 06 06:02:39 crc kubenswrapper[4957]: E1206 06:02:39.767351 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 06 06:02:39 crc kubenswrapper[4957]: E1206 06:02:39.767518 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4hcvj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-f45dr_openstack(7d3a3efb-65b9-4a90-bded-99969b537744): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:02:39 crc kubenswrapper[4957]: E1206 06:02:39.769072 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-f45dr" podUID="7d3a3efb-65b9-4a90-bded-99969b537744" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.858062 4957 scope.go:117] "RemoveContainer" containerID="33b22b04c08c8a56fe1ebe806b229b124fcebfc0fe4f405e0c3ae983be758002" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.874041 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:39 crc kubenswrapper[4957]: I1206 06:02:39.961997 4957 scope.go:117] "RemoveContainer" containerID="94c98f2b707f8620623fbd278850e7ffb3a28d7a1d5786660ad3c897c2fb83e1" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.026236 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-config\") pod \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.026621 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-combined-ca-bundle\") pod \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.026842 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd4qj\" (UniqueName: \"kubernetes.io/projected/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-kube-api-access-qd4qj\") pod \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\" (UID: \"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc\") " Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.032345 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-kube-api-access-qd4qj" (OuterVolumeSpecName: "kube-api-access-qd4qj") pod "d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" (UID: "d90cadb8-4faf-4c1a-b198-0a6b4d537bdc"). InnerVolumeSpecName "kube-api-access-qd4qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.057999 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-config" (OuterVolumeSpecName: "config") pod "d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" (UID: "d90cadb8-4faf-4c1a-b198-0a6b4d537bdc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.074108 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" (UID: "d90cadb8-4faf-4c1a-b198-0a6b4d537bdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.128979 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.129022 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.129040 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd4qj\" (UniqueName: \"kubernetes.io/projected/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc-kube-api-access-qd4qj\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.182365 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.182432 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.182479 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.183310 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3b7a79d60429782e416177d4e1df59e205a0e6bbe183ad64871d22be42eb374"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.183377 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://b3b7a79d60429782e416177d4e1df59e205a0e6bbe183ad64871d22be42eb374" gracePeriod=600 Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.218257 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76c6cdbd84-nmkc9"] Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.267914 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-xl9rr" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.340936 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="b3b7a79d60429782e416177d4e1df59e205a0e6bbe183ad64871d22be42eb374" exitCode=0 Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.341016 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"b3b7a79d60429782e416177d4e1df59e205a0e6bbe183ad64871d22be42eb374"} Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.344445 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gv9kk" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.344633 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gv9kk" event={"ID":"d90cadb8-4faf-4c1a-b198-0a6b4d537bdc","Type":"ContainerDied","Data":"f891d3bac3b735e01b8c1f10d2a06a52509554b1b9ecdc018113e1d3d7c334a4"} Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.344687 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f891d3bac3b735e01b8c1f10d2a06a52509554b1b9ecdc018113e1d3d7c334a4" Dec 06 06:02:40 crc kubenswrapper[4957]: E1206 06:02:40.346293 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-f45dr" podUID="7d3a3efb-65b9-4a90-bded-99969b537744" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.369515 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c76d956-bsr6l"] Dec 06 06:02:40 crc kubenswrapper[4957]: W1206 06:02:40.444147 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d4332b6_8252_435b_9a1c_bc3f27addaea.slice/crio-f62c753849f8e5f33d057b0da12e2d4917abe7a5d94603dfcced8744ef0f6e41 WatchSource:0}: Error finding container f62c753849f8e5f33d057b0da12e2d4917abe7a5d94603dfcced8744ef0f6e41: Status 404 returned error can't find the container with id f62c753849f8e5f33d057b0da12e2d4917abe7a5d94603dfcced8744ef0f6e41 Dec 06 06:02:40 crc kubenswrapper[4957]: W1206 06:02:40.447787 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb06a82d_1610_4985_8168_c95f32ad2ff6.slice/crio-ca191c54371fbcafc9be269f74639f33d8b913b2f9af546b92bf472fe17cd647 WatchSource:0}: Error finding container ca191c54371fbcafc9be269f74639f33d8b913b2f9af546b92bf472fe17cd647: Status 404 returned error can't find the container with id ca191c54371fbcafc9be269f74639f33d8b913b2f9af546b92bf472fe17cd647 Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.460566 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.463083 4957 scope.go:117] "RemoveContainer" containerID="8d9885a314fd8362aab5bd731604c4e65aee6a5e03ad4fb94df9824b6245b31f" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.493280 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bqh2v"] Dec 06 06:02:40 crc kubenswrapper[4957]: W1206 06:02:40.497521 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod376b5f5c_4e23_4e75_9619_698a4d95eab5.slice/crio-dd745fa12b620cbfc218082405935e489c68f692fa141055e3118c2aba3a007a WatchSource:0}: Error finding container dd745fa12b620cbfc218082405935e489c68f692fa141055e3118c2aba3a007a: Status 404 returned error can't find the container with id dd745fa12b620cbfc218082405935e489c68f692fa141055e3118c2aba3a007a Dec 06 06:02:40 crc kubenswrapper[4957]: W1206 06:02:40.505135 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8a45479_cc0d_40ce_b54d_d65ee56a6c36.slice/crio-32c49e5c1dce4dd51a59b5749a83e6f74bfcadff307b13e4a460199a76e264b0 WatchSource:0}: Error finding container 32c49e5c1dce4dd51a59b5749a83e6f74bfcadff307b13e4a460199a76e264b0: Status 404 returned error can't find the container with id 32c49e5c1dce4dd51a59b5749a83e6f74bfcadff307b13e4a460199a76e264b0 Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.632344 4957 scope.go:117] "RemoveContainer" containerID="f3f75ddbe4ecf31b2a5519fb70a8ed4815614284758f53e65c5fc0533e95c0a3" Dec 06 06:02:40 crc kubenswrapper[4957]: I1206 06:02:40.674446 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c64c7f6-187a-43dc-9994-20efbe94411e" path="/var/lib/kubelet/pods/8c64c7f6-187a-43dc-9994-20efbe94411e/volumes" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.156116 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zls57"] Dec 06 06:02:41 crc kubenswrapper[4957]: E1206 06:02:41.156726 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" containerName="neutron-db-sync" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.156740 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" containerName="neutron-db-sync" Dec 06 06:02:41 crc kubenswrapper[4957]: E1206 06:02:41.156770 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="init" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.156776 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="init" Dec 06 06:02:41 crc kubenswrapper[4957]: E1206 06:02:41.156792 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="dnsmasq-dns" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.156798 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="dnsmasq-dns" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.156975 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" containerName="neutron-db-sync" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.156991 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="06b68237-df64-484c-859a-9dcb92a5059a" containerName="dnsmasq-dns" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.162214 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.173610 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zls57"] Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.267600 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8xhn\" (UniqueName: \"kubernetes.io/projected/7859baf4-005a-4dad-8670-5d0ddc28c6e9-kube-api-access-h8xhn\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.267658 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.267698 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.267747 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-svc\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.267814 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-config\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.267903 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.289271 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-547bff997b-g5msk"] Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.291349 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.297132 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-srkwj" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.297371 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.297522 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.297683 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.312503 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-547bff997b-g5msk"] Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371211 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371342 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8xhn\" (UniqueName: \"kubernetes.io/projected/7859baf4-005a-4dad-8670-5d0ddc28c6e9-kube-api-access-h8xhn\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371680 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-httpd-config\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371726 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371762 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57cn8\" (UniqueName: \"kubernetes.io/projected/3050ec84-871a-460d-b50b-e52c87732f4d-kube-api-access-57cn8\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371800 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371888 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-ovndb-tls-certs\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.371965 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-svc\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.372042 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-config\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.372135 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-config\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.372168 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-combined-ca-bundle\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.372438 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.372647 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.372897 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.373209 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-config\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.374163 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-svc\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.379768 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c76d956-bsr6l" event={"ID":"cb06a82d-1610-4985-8168-c95f32ad2ff6","Type":"ContainerStarted","Data":"708f4f08f381616eb446f1d55a59cdc1f1c282b7b39d46501a2831435bea2768"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.379844 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c76d956-bsr6l" event={"ID":"cb06a82d-1610-4985-8168-c95f32ad2ff6","Type":"ContainerStarted","Data":"ca191c54371fbcafc9be269f74639f33d8b913b2f9af546b92bf472fe17cd647"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.382796 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c6cdbd84-nmkc9" event={"ID":"3d4332b6-8252-435b-9a1c-bc3f27addaea","Type":"ContainerStarted","Data":"1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.382858 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c6cdbd84-nmkc9" event={"ID":"3d4332b6-8252-435b-9a1c-bc3f27addaea","Type":"ContainerStarted","Data":"f62c753849f8e5f33d057b0da12e2d4917abe7a5d94603dfcced8744ef0f6e41"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.385875 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a0cdab5-b096-405a-9018-9319eccd9b9f","Type":"ContainerStarted","Data":"f0629cc1dd119ea5eda64558a77e2893c31de9d4db835e9ce7d893477706c4e5"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.386051 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-log" containerID="cri-o://92947cccf0669f5e9c7584d557033334706aa9b8d23cc7787538bea6adb644fa" gracePeriod=30 Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.386674 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-httpd" containerID="cri-o://f0629cc1dd119ea5eda64558a77e2893c31de9d4db835e9ce7d893477706c4e5" gracePeriod=30 Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.388887 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8xhn\" (UniqueName: \"kubernetes.io/projected/7859baf4-005a-4dad-8670-5d0ddc28c6e9-kube-api-access-h8xhn\") pod \"dnsmasq-dns-55f844cf75-zls57\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.398271 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bqh2v" event={"ID":"e8a45479-cc0d-40ce-b54d-d65ee56a6c36","Type":"ContainerStarted","Data":"67c4ebc25385b97d5d798360d85799ca98f251b184a5b874655814f3da6cd083"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.398315 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bqh2v" event={"ID":"e8a45479-cc0d-40ce-b54d-d65ee56a6c36","Type":"ContainerStarted","Data":"32c49e5c1dce4dd51a59b5749a83e6f74bfcadff307b13e4a460199a76e264b0"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.406215 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.414825 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=32.414804709 podStartE2EDuration="32.414804709s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:41.407409039 +0000 UTC m=+1401.057676691" watchObservedRunningTime="2025-12-06 06:02:41.414804709 +0000 UTC m=+1401.065072341" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.420362 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerStarted","Data":"11df35ebea1e26e949fe4e06981f15b1f30286868b24ce1b74fd58e1315a6fd6"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.422603 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54c4596cfc-p2m5g" event={"ID":"90caf120-13de-4d6f-9e95-2954a448e874","Type":"ContainerStarted","Data":"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.426337 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"376b5f5c-4e23-4e75-9619-698a4d95eab5","Type":"ContainerStarted","Data":"dd745fa12b620cbfc218082405935e489c68f692fa141055e3118c2aba3a007a"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.435419 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd755cc-4mppd" event={"ID":"430ef178-b9c0-4759-979b-92620ceca29f","Type":"ContainerStarted","Data":"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933"} Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.450503 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bqh2v" podStartSLOduration=5.450487479 podStartE2EDuration="5.450487479s" podCreationTimestamp="2025-12-06 06:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:41.44384797 +0000 UTC m=+1401.094115622" watchObservedRunningTime="2025-12-06 06:02:41.450487479 +0000 UTC m=+1401.100755111" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.473355 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-httpd-config\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.473443 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57cn8\" (UniqueName: \"kubernetes.io/projected/3050ec84-871a-460d-b50b-e52c87732f4d-kube-api-access-57cn8\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.473480 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-ovndb-tls-certs\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.477090 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-config\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.477198 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-combined-ca-bundle\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.489133 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-httpd-config\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.492902 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-ovndb-tls-certs\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.495557 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-config\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.501490 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-combined-ca-bundle\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.516776 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57cn8\" (UniqueName: \"kubernetes.io/projected/3050ec84-871a-460d-b50b-e52c87732f4d-kube-api-access-57cn8\") pod \"neutron-547bff997b-g5msk\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.553542 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:41 crc kubenswrapper[4957]: I1206 06:02:41.584738 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:41 crc kubenswrapper[4957]: E1206 06:02:41.930051 4957 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a0cdab5_b096_405a_9018_9319eccd9b9f.slice/crio-conmon-f0629cc1dd119ea5eda64558a77e2893c31de9d4db835e9ce7d893477706c4e5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a0cdab5_b096_405a_9018_9319eccd9b9f.slice/crio-f0629cc1dd119ea5eda64558a77e2893c31de9d4db835e9ce7d893477706c4e5.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.195130 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zls57"] Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.362680 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-547bff997b-g5msk"] Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.465902 4957 generic.go:334] "Generic (PLEG): container finished" podID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerID="f0629cc1dd119ea5eda64558a77e2893c31de9d4db835e9ce7d893477706c4e5" exitCode=0 Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.466122 4957 generic.go:334] "Generic (PLEG): container finished" podID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerID="92947cccf0669f5e9c7584d557033334706aa9b8d23cc7787538bea6adb644fa" exitCode=143 Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.466216 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a0cdab5-b096-405a-9018-9319eccd9b9f","Type":"ContainerDied","Data":"f0629cc1dd119ea5eda64558a77e2893c31de9d4db835e9ce7d893477706c4e5"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.466264 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a0cdab5-b096-405a-9018-9319eccd9b9f","Type":"ContainerDied","Data":"92947cccf0669f5e9c7584d557033334706aa9b8d23cc7787538bea6adb644fa"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.466277 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5a0cdab5-b096-405a-9018-9319eccd9b9f","Type":"ContainerDied","Data":"97752f2c9e232dbbcb18c4553b6d4731971cbb1b74af236badd77745f11eb59c"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.466289 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97752f2c9e232dbbcb18c4553b6d4731971cbb1b74af236badd77745f11eb59c" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.471687 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zls57" event={"ID":"7859baf4-005a-4dad-8670-5d0ddc28c6e9","Type":"ContainerStarted","Data":"cea40f5f52a4243ac0e166617d275b12f35bbd8592313be1a8f09507c347a9f1"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.483171 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54c4596cfc-p2m5g" event={"ID":"90caf120-13de-4d6f-9e95-2954a448e874","Type":"ContainerStarted","Data":"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.483278 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-54c4596cfc-p2m5g" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon-log" containerID="cri-o://fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658" gracePeriod=30 Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.483395 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-54c4596cfc-p2m5g" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon" containerID="cri-o://a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff" gracePeriod=30 Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.497232 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.497572 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"376b5f5c-4e23-4e75-9619-698a4d95eab5","Type":"ContainerStarted","Data":"85679ee7f099f70cfabd5b1d2aeb68997468664d5994775abff132ffa150a703"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.509278 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-54c4596cfc-p2m5g" podStartSLOduration=4.666732545 podStartE2EDuration="31.509259876s" podCreationTimestamp="2025-12-06 06:02:11 +0000 UTC" firstStartedPulling="2025-12-06 06:02:12.9250207 +0000 UTC m=+1372.575288332" lastFinishedPulling="2025-12-06 06:02:39.767548001 +0000 UTC m=+1399.417815663" observedRunningTime="2025-12-06 06:02:42.505166136 +0000 UTC m=+1402.155433768" watchObservedRunningTime="2025-12-06 06:02:42.509259876 +0000 UTC m=+1402.159527498" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.514754 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c76d956-bsr6l" event={"ID":"cb06a82d-1610-4985-8168-c95f32ad2ff6","Type":"ContainerStarted","Data":"145540ebdb3c7430fbc4148230e575a1f9ec7cdc487d48ec07b08b8ccfde5582"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.519047 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd755cc-4mppd" event={"ID":"430ef178-b9c0-4759-979b-92620ceca29f","Type":"ContainerStarted","Data":"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.519183 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-bcdd755cc-4mppd" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon-log" containerID="cri-o://df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933" gracePeriod=30 Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.519226 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-bcdd755cc-4mppd" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon" containerID="cri-o://2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0" gracePeriod=30 Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.525766 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c6cdbd84-nmkc9" event={"ID":"3d4332b6-8252-435b-9a1c-bc3f27addaea","Type":"ContainerStarted","Data":"da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.540403 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-547bff997b-g5msk" event={"ID":"3050ec84-871a-460d-b50b-e52c87732f4d","Type":"ContainerStarted","Data":"c41136494c1847077362e62b4e6eff212bf68f3610022d04ae3529c9e66727f2"} Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.555529 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5c76d956-bsr6l" podStartSLOduration=24.555503622 podStartE2EDuration="24.555503622s" podCreationTimestamp="2025-12-06 06:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:42.550108567 +0000 UTC m=+1402.200376219" watchObservedRunningTime="2025-12-06 06:02:42.555503622 +0000 UTC m=+1402.205771264" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.682198 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-bcdd755cc-4mppd" podStartSLOduration=6.404297277 podStartE2EDuration="33.674559248s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="2025-12-06 06:02:12.62835256 +0000 UTC m=+1372.278620192" lastFinishedPulling="2025-12-06 06:02:39.898614531 +0000 UTC m=+1399.548882163" observedRunningTime="2025-12-06 06:02:42.646362609 +0000 UTC m=+1402.296630251" watchObservedRunningTime="2025-12-06 06:02:42.674559248 +0000 UTC m=+1402.324826880" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712402 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-scripts\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712471 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-internal-tls-certs\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712555 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-76c6cdbd84-nmkc9" podStartSLOduration=24.712530141 podStartE2EDuration="24.712530141s" podCreationTimestamp="2025-12-06 06:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:42.682733159 +0000 UTC m=+1402.333000871" watchObservedRunningTime="2025-12-06 06:02:42.712530141 +0000 UTC m=+1402.362797773" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712658 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-config-data\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712715 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712744 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-httpd-run\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712789 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-logs\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712871 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6pbm\" (UniqueName: \"kubernetes.io/projected/5a0cdab5-b096-405a-9018-9319eccd9b9f-kube-api-access-s6pbm\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.712916 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-combined-ca-bundle\") pod \"5a0cdab5-b096-405a-9018-9319eccd9b9f\" (UID: \"5a0cdab5-b096-405a-9018-9319eccd9b9f\") " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.713342 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.713512 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-logs" (OuterVolumeSpecName: "logs") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.724354 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-scripts" (OuterVolumeSpecName: "scripts") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.729179 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a0cdab5-b096-405a-9018-9319eccd9b9f-kube-api-access-s6pbm" (OuterVolumeSpecName: "kube-api-access-s6pbm") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "kube-api-access-s6pbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.732653 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.812123 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.816568 4957 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.816598 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a0cdab5-b096-405a-9018-9319eccd9b9f-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.816607 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6pbm\" (UniqueName: \"kubernetes.io/projected/5a0cdab5-b096-405a-9018-9319eccd9b9f-kube-api-access-s6pbm\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.816647 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.816657 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.816677 4957 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.860878 4957 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.863057 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.867382 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-config-data" (OuterVolumeSpecName: "config-data") pod "5a0cdab5-b096-405a-9018-9319eccd9b9f" (UID: "5a0cdab5-b096-405a-9018-9319eccd9b9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.917919 4957 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.917948 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0cdab5-b096-405a-9018-9319eccd9b9f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:42 crc kubenswrapper[4957]: I1206 06:02:42.917958 4957 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.571969 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-547bff997b-g5msk" event={"ID":"3050ec84-871a-460d-b50b-e52c87732f4d","Type":"ContainerStarted","Data":"9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f"} Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.572632 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-547bff997b-g5msk" event={"ID":"3050ec84-871a-460d-b50b-e52c87732f4d","Type":"ContainerStarted","Data":"5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00"} Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.572934 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.584997 4957 generic.go:334] "Generic (PLEG): container finished" podID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerID="93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6" exitCode=0 Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.585293 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zls57" event={"ID":"7859baf4-005a-4dad-8670-5d0ddc28c6e9","Type":"ContainerDied","Data":"93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6"} Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.600724 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-547bff997b-g5msk" podStartSLOduration=2.600702693 podStartE2EDuration="2.600702693s" podCreationTimestamp="2025-12-06 06:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:43.592116752 +0000 UTC m=+1403.242384414" watchObservedRunningTime="2025-12-06 06:02:43.600702693 +0000 UTC m=+1403.250970325" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.645697 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.650021 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"376b5f5c-4e23-4e75-9619-698a4d95eab5","Type":"ContainerStarted","Data":"0a057a094f04b6b804a78b44a8c9aa245162822d6f56d0aa83222484239b62ea"} Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.711286 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.711269851 podStartE2EDuration="7.711269851s" podCreationTimestamp="2025-12-06 06:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:43.694704616 +0000 UTC m=+1403.344972248" watchObservedRunningTime="2025-12-06 06:02:43.711269851 +0000 UTC m=+1403.361537483" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.786668 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.802246 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.815419 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:43 crc kubenswrapper[4957]: E1206 06:02:43.815914 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-httpd" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.815934 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-httpd" Dec 06 06:02:43 crc kubenswrapper[4957]: E1206 06:02:43.815956 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-log" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.815964 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-log" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.816207 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-httpd" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.816223 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" containerName="glance-log" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.817423 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.820237 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.820535 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.861536 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943534 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943600 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943671 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943699 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943721 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g46t9\" (UniqueName: \"kubernetes.io/projected/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-kube-api-access-g46t9\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943742 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943761 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:43 crc kubenswrapper[4957]: I1206 06:02:43.943795 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.011440 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-575d75c97c-wblw4"] Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.013826 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.017002 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.017121 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.045546 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-575d75c97c-wblw4"] Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046482 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046559 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046646 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046683 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046712 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g46t9\" (UniqueName: \"kubernetes.io/projected/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-kube-api-access-g46t9\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046752 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046776 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.046815 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.050308 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.050320 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.055110 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.056279 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.063995 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.064402 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.072322 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.072578 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g46t9\" (UniqueName: \"kubernetes.io/projected/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-kube-api-access-g46t9\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.095244 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.145790 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.148755 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-public-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.148813 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-ovndb-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.148870 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-combined-ca-bundle\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.148907 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znx76\" (UniqueName: \"kubernetes.io/projected/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-kube-api-access-znx76\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.149101 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-config\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.149177 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-internal-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.149289 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-httpd-config\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.257262 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-internal-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.257739 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-httpd-config\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.257890 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-public-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.257948 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-ovndb-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.258028 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-combined-ca-bundle\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.258095 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znx76\" (UniqueName: \"kubernetes.io/projected/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-kube-api-access-znx76\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.258162 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-config\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.262608 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-config\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.265417 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-httpd-config\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.268970 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-internal-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.269010 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-ovndb-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.272455 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-combined-ca-bundle\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.273250 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-public-tls-certs\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.278414 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znx76\" (UniqueName: \"kubernetes.io/projected/a7c5d2f8-fce6-4689-8090-9ccbc8bb795e-kube-api-access-znx76\") pod \"neutron-575d75c97c-wblw4\" (UID: \"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e\") " pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.330260 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.693298 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a0cdab5-b096-405a-9018-9319eccd9b9f" path="/var/lib/kubelet/pods/5a0cdab5-b096-405a-9018-9319eccd9b9f/volumes" Dec 06 06:02:44 crc kubenswrapper[4957]: I1206 06:02:44.964400 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-575d75c97c-wblw4"] Dec 06 06:02:44 crc kubenswrapper[4957]: W1206 06:02:44.982052 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7c5d2f8_fce6_4689_8090_9ccbc8bb795e.slice/crio-c408f20bec1dcc2f56b2768f06224516302d41322a221f2351112ec1591c4c0f WatchSource:0}: Error finding container c408f20bec1dcc2f56b2768f06224516302d41322a221f2351112ec1591c4c0f: Status 404 returned error can't find the container with id c408f20bec1dcc2f56b2768f06224516302d41322a221f2351112ec1591c4c0f Dec 06 06:02:45 crc kubenswrapper[4957]: I1206 06:02:45.488622 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:02:45 crc kubenswrapper[4957]: W1206 06:02:45.491346 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1085fe7_6d61_4e23_85ba_51ec7740b8c4.slice/crio-87c72743a2df1a51d9015cfede0aee1c76428c106ec7d5fe729d4c01c8f06d2c WatchSource:0}: Error finding container 87c72743a2df1a51d9015cfede0aee1c76428c106ec7d5fe729d4c01c8f06d2c: Status 404 returned error can't find the container with id 87c72743a2df1a51d9015cfede0aee1c76428c106ec7d5fe729d4c01c8f06d2c Dec 06 06:02:45 crc kubenswrapper[4957]: I1206 06:02:45.675901 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-575d75c97c-wblw4" event={"ID":"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e","Type":"ContainerStarted","Data":"c408f20bec1dcc2f56b2768f06224516302d41322a221f2351112ec1591c4c0f"} Dec 06 06:02:45 crc kubenswrapper[4957]: I1206 06:02:45.681671 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1085fe7-6d61-4e23-85ba-51ec7740b8c4","Type":"ContainerStarted","Data":"87c72743a2df1a51d9015cfede0aee1c76428c106ec7d5fe729d4c01c8f06d2c"} Dec 06 06:02:46 crc kubenswrapper[4957]: I1206 06:02:46.754014 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:02:46 crc kubenswrapper[4957]: I1206 06:02:46.754396 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:02:46 crc kubenswrapper[4957]: I1206 06:02:46.790443 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:02:46 crc kubenswrapper[4957]: I1206 06:02:46.797213 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:02:47 crc kubenswrapper[4957]: I1206 06:02:47.711081 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:02:47 crc kubenswrapper[4957]: I1206 06:02:47.711896 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:02:48 crc kubenswrapper[4957]: I1206 06:02:48.723174 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zls57" event={"ID":"7859baf4-005a-4dad-8670-5d0ddc28c6e9","Type":"ContainerStarted","Data":"2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986"} Dec 06 06:02:48 crc kubenswrapper[4957]: I1206 06:02:48.723503 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:48 crc kubenswrapper[4957]: I1206 06:02:48.725922 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-575d75c97c-wblw4" event={"ID":"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e","Type":"ContainerStarted","Data":"a1f4e0220f4501cbe9b6e271fb19e4f97c8588a23bc5ca17974e3afbe9ce99e6"} Dec 06 06:02:48 crc kubenswrapper[4957]: I1206 06:02:48.746822 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-zls57" podStartSLOduration=7.746800749 podStartE2EDuration="7.746800749s" podCreationTimestamp="2025-12-06 06:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:48.745441213 +0000 UTC m=+1408.395708855" watchObservedRunningTime="2025-12-06 06:02:48.746800749 +0000 UTC m=+1408.397068381" Dec 06 06:02:48 crc kubenswrapper[4957]: I1206 06:02:48.788866 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:48 crc kubenswrapper[4957]: I1206 06:02:48.789910 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:02:49 crc kubenswrapper[4957]: I1206 06:02:49.091444 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:49 crc kubenswrapper[4957]: I1206 06:02:49.091535 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:02:49 crc kubenswrapper[4957]: I1206 06:02:49.914673 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:02:50 crc kubenswrapper[4957]: I1206 06:02:50.172479 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:02:50 crc kubenswrapper[4957]: I1206 06:02:50.746251 4957 generic.go:334] "Generic (PLEG): container finished" podID="e8a45479-cc0d-40ce-b54d-d65ee56a6c36" containerID="67c4ebc25385b97d5d798360d85799ca98f251b184a5b874655814f3da6cd083" exitCode=0 Dec 06 06:02:50 crc kubenswrapper[4957]: I1206 06:02:50.746301 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bqh2v" event={"ID":"e8a45479-cc0d-40ce-b54d-d65ee56a6c36","Type":"ContainerDied","Data":"67c4ebc25385b97d5d798360d85799ca98f251b184a5b874655814f3da6cd083"} Dec 06 06:02:50 crc kubenswrapper[4957]: I1206 06:02:50.867066 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:02:52 crc kubenswrapper[4957]: I1206 06:02:52.112263 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.324351 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.459419 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-config-data\") pod \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.459487 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq9dn\" (UniqueName: \"kubernetes.io/projected/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-kube-api-access-pq9dn\") pod \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.459659 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-scripts\") pod \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.459736 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-combined-ca-bundle\") pod \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.459800 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-fernet-keys\") pod \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.459861 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-credential-keys\") pod \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\" (UID: \"e8a45479-cc0d-40ce-b54d-d65ee56a6c36\") " Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.469011 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e8a45479-cc0d-40ce-b54d-d65ee56a6c36" (UID: "e8a45479-cc0d-40ce-b54d-d65ee56a6c36"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.469083 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-kube-api-access-pq9dn" (OuterVolumeSpecName: "kube-api-access-pq9dn") pod "e8a45479-cc0d-40ce-b54d-d65ee56a6c36" (UID: "e8a45479-cc0d-40ce-b54d-d65ee56a6c36"). InnerVolumeSpecName "kube-api-access-pq9dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.469195 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e8a45479-cc0d-40ce-b54d-d65ee56a6c36" (UID: "e8a45479-cc0d-40ce-b54d-d65ee56a6c36"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.475339 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-scripts" (OuterVolumeSpecName: "scripts") pod "e8a45479-cc0d-40ce-b54d-d65ee56a6c36" (UID: "e8a45479-cc0d-40ce-b54d-d65ee56a6c36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.499020 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-config-data" (OuterVolumeSpecName: "config-data") pod "e8a45479-cc0d-40ce-b54d-d65ee56a6c36" (UID: "e8a45479-cc0d-40ce-b54d-d65ee56a6c36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.499031 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8a45479-cc0d-40ce-b54d-d65ee56a6c36" (UID: "e8a45479-cc0d-40ce-b54d-d65ee56a6c36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.565881 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.565911 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.565923 4957 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.565932 4957 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.565941 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.565949 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq9dn\" (UniqueName: \"kubernetes.io/projected/e8a45479-cc0d-40ce-b54d-d65ee56a6c36-kube-api-access-pq9dn\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.787908 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bqh2v" event={"ID":"e8a45479-cc0d-40ce-b54d-d65ee56a6c36","Type":"ContainerDied","Data":"32c49e5c1dce4dd51a59b5749a83e6f74bfcadff307b13e4a460199a76e264b0"} Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.788184 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32c49e5c1dce4dd51a59b5749a83e6f74bfcadff307b13e4a460199a76e264b0" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.788638 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bqh2v" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.793209 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1085fe7-6d61-4e23-85ba-51ec7740b8c4","Type":"ContainerStarted","Data":"4fe6ed5c1695b975e31be44f6177795022263bc46dadc34ae744b6e40adb918d"} Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.795331 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerStarted","Data":"77234a9a9030cd0d0b92e72149b99781c3b5e68217509a6458ccbebbc9ff5c84"} Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.797001 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-575d75c97c-wblw4" event={"ID":"a7c5d2f8-fce6-4689-8090-9ccbc8bb795e","Type":"ContainerStarted","Data":"50aab26e7b02bbc9fed7d6e018374785528a04544e85a48e71c6f872688796ea"} Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.797160 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.799130 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zsklh" event={"ID":"c167b57e-b623-4080-a770-d87233e00128","Type":"ContainerStarted","Data":"cea7f2dbbfccaccec7e7a87cd579143a35051da4c050e15ff8d8782c1f013fba"} Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.843050 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-575d75c97c-wblw4" podStartSLOduration=10.843016631 podStartE2EDuration="10.843016631s" podCreationTimestamp="2025-12-06 06:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:53.813374354 +0000 UTC m=+1413.463642006" watchObservedRunningTime="2025-12-06 06:02:53.843016631 +0000 UTC m=+1413.493284263" Dec 06 06:02:53 crc kubenswrapper[4957]: I1206 06:02:53.872019 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zsklh" podStartSLOduration=4.479798153 podStartE2EDuration="44.871995513s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="2025-12-06 06:02:12.696560956 +0000 UTC m=+1372.346828588" lastFinishedPulling="2025-12-06 06:02:53.088758306 +0000 UTC m=+1412.739025948" observedRunningTime="2025-12-06 06:02:53.831726047 +0000 UTC m=+1413.481993699" watchObservedRunningTime="2025-12-06 06:02:53.871995513 +0000 UTC m=+1413.522263145" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.420656 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7c7fdbc8b-njnq6"] Dec 06 06:02:54 crc kubenswrapper[4957]: E1206 06:02:54.421295 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a45479-cc0d-40ce-b54d-d65ee56a6c36" containerName="keystone-bootstrap" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.421311 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a45479-cc0d-40ce-b54d-d65ee56a6c36" containerName="keystone-bootstrap" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.421501 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8a45479-cc0d-40ce-b54d-d65ee56a6c36" containerName="keystone-bootstrap" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.422077 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.424875 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.425041 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.425168 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.425292 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.425396 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.426036 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-hx52c" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.430038 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c7fdbc8b-njnq6"] Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.481885 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-public-tls-certs\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.481935 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-internal-tls-certs\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.481958 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-config-data\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.481982 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-scripts\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.482000 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-credential-keys\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.482044 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-fernet-keys\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.482070 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-combined-ca-bundle\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.482149 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8lq6\" (UniqueName: \"kubernetes.io/projected/95511bb1-a0e3-4761-aba1-fd10b37df59f-kube-api-access-b8lq6\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584109 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-public-tls-certs\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584162 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-internal-tls-certs\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584186 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-config-data\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584211 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-scripts\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584232 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-credential-keys\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584280 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-fernet-keys\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584308 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-combined-ca-bundle\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.584341 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8lq6\" (UniqueName: \"kubernetes.io/projected/95511bb1-a0e3-4761-aba1-fd10b37df59f-kube-api-access-b8lq6\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.591678 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-credential-keys\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.591706 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-public-tls-certs\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.591946 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-scripts\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.593745 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-combined-ca-bundle\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.594518 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-config-data\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.598241 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-internal-tls-certs\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.612616 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8lq6\" (UniqueName: \"kubernetes.io/projected/95511bb1-a0e3-4761-aba1-fd10b37df59f-kube-api-access-b8lq6\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.614739 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95511bb1-a0e3-4761-aba1-fd10b37df59f-fernet-keys\") pod \"keystone-7c7fdbc8b-njnq6\" (UID: \"95511bb1-a0e3-4761-aba1-fd10b37df59f\") " pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.738250 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.840245 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6szxq" event={"ID":"caa7934f-ce05-4cac-b198-b67ffa3e96ed","Type":"ContainerStarted","Data":"b70e91a0cf14b35a501421be0dec610a6bfe3f59df20f9c4f06da3e3c089f43d"} Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.846544 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1085fe7-6d61-4e23-85ba-51ec7740b8c4","Type":"ContainerStarted","Data":"54444629802b56358878379ddd680c333de89593fda09553e8fe24d76d03288b"} Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.882093 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-6szxq" podStartSLOduration=3.269797781 podStartE2EDuration="45.882069517s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="2025-12-06 06:02:10.506750726 +0000 UTC m=+1370.157018358" lastFinishedPulling="2025-12-06 06:02:53.119022462 +0000 UTC m=+1412.769290094" observedRunningTime="2025-12-06 06:02:54.853851447 +0000 UTC m=+1414.504119079" watchObservedRunningTime="2025-12-06 06:02:54.882069517 +0000 UTC m=+1414.532337149" Dec 06 06:02:54 crc kubenswrapper[4957]: I1206 06:02:54.897608 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.897594766 podStartE2EDuration="11.897594766s" podCreationTimestamp="2025-12-06 06:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:54.877751292 +0000 UTC m=+1414.528018944" watchObservedRunningTime="2025-12-06 06:02:54.897594766 +0000 UTC m=+1414.547862398" Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.233954 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c7fdbc8b-njnq6"] Dec 06 06:02:55 crc kubenswrapper[4957]: W1206 06:02:55.265239 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95511bb1_a0e3_4761_aba1_fd10b37df59f.slice/crio-25d5275f1545ab249ee5ecb81353588b9697f382d0f250e011d23fce0d5ae7a0 WatchSource:0}: Error finding container 25d5275f1545ab249ee5ecb81353588b9697f382d0f250e011d23fce0d5ae7a0: Status 404 returned error can't find the container with id 25d5275f1545ab249ee5ecb81353588b9697f382d0f250e011d23fce0d5ae7a0 Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.856131 4957 generic.go:334] "Generic (PLEG): container finished" podID="c167b57e-b623-4080-a770-d87233e00128" containerID="cea7f2dbbfccaccec7e7a87cd579143a35051da4c050e15ff8d8782c1f013fba" exitCode=0 Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.856381 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zsklh" event={"ID":"c167b57e-b623-4080-a770-d87233e00128","Type":"ContainerDied","Data":"cea7f2dbbfccaccec7e7a87cd579143a35051da4c050e15ff8d8782c1f013fba"} Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.860088 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f45dr" event={"ID":"7d3a3efb-65b9-4a90-bded-99969b537744","Type":"ContainerStarted","Data":"14b61a9a5561fe1188d263899b81f4fc8d0af35a3867fae10c46438c1746f01a"} Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.863761 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c7fdbc8b-njnq6" event={"ID":"95511bb1-a0e3-4761-aba1-fd10b37df59f","Type":"ContainerStarted","Data":"531b5d03063ac5134702bd6c587c8f5cae34b55423ac795e8b0055328a57a2af"} Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.863804 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c7fdbc8b-njnq6" event={"ID":"95511bb1-a0e3-4761-aba1-fd10b37df59f","Type":"ContainerStarted","Data":"25d5275f1545ab249ee5ecb81353588b9697f382d0f250e011d23fce0d5ae7a0"} Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.922644 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7c7fdbc8b-njnq6" podStartSLOduration=1.9226231839999999 podStartE2EDuration="1.922623184s" podCreationTimestamp="2025-12-06 06:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:02:55.901246638 +0000 UTC m=+1415.551514290" watchObservedRunningTime="2025-12-06 06:02:55.922623184 +0000 UTC m=+1415.572890816" Dec 06 06:02:55 crc kubenswrapper[4957]: I1206 06:02:55.922943 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-f45dr" podStartSLOduration=4.544635428 podStartE2EDuration="46.922939783s" podCreationTimestamp="2025-12-06 06:02:09 +0000 UTC" firstStartedPulling="2025-12-06 06:02:12.727948051 +0000 UTC m=+1372.378215673" lastFinishedPulling="2025-12-06 06:02:55.106252396 +0000 UTC m=+1414.756520028" observedRunningTime="2025-12-06 06:02:55.912653706 +0000 UTC m=+1415.562921338" watchObservedRunningTime="2025-12-06 06:02:55.922939783 +0000 UTC m=+1415.573207415" Dec 06 06:02:56 crc kubenswrapper[4957]: I1206 06:02:56.555033 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:02:56 crc kubenswrapper[4957]: I1206 06:02:56.657217 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qmvb2"] Dec 06 06:02:56 crc kubenswrapper[4957]: I1206 06:02:56.661642 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" podUID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerName="dnsmasq-dns" containerID="cri-o://0d35b8b3da1aab99f8dc72cf460458a1afe19bd0cdc7862664f55a885a4bbc56" gracePeriod=10 Dec 06 06:02:56 crc kubenswrapper[4957]: I1206 06:02:56.908804 4957 generic.go:334] "Generic (PLEG): container finished" podID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerID="0d35b8b3da1aab99f8dc72cf460458a1afe19bd0cdc7862664f55a885a4bbc56" exitCode=0 Dec 06 06:02:56 crc kubenswrapper[4957]: I1206 06:02:56.909140 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" event={"ID":"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e","Type":"ContainerDied","Data":"0d35b8b3da1aab99f8dc72cf460458a1afe19bd0cdc7862664f55a885a4bbc56"} Dec 06 06:02:56 crc kubenswrapper[4957]: I1206 06:02:56.910276 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.245540 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.359266 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-svc\") pod \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.359322 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-config\") pod \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.359367 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-swift-storage-0\") pod \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.359404 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv2vv\" (UniqueName: \"kubernetes.io/projected/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-kube-api-access-hv2vv\") pod \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.359468 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-nb\") pod \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.359532 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb\") pod \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.370761 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.378938 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-kube-api-access-hv2vv" (OuterVolumeSpecName: "kube-api-access-hv2vv") pod "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" (UID: "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e"). InnerVolumeSpecName "kube-api-access-hv2vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.422055 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-config" (OuterVolumeSpecName: "config") pod "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" (UID: "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.424147 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" (UID: "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.454585 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" (UID: "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.461322 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" (UID: "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.461340 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-config-data\") pod \"c167b57e-b623-4080-a770-d87233e00128\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.461426 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6td2m\" (UniqueName: \"kubernetes.io/projected/c167b57e-b623-4080-a770-d87233e00128-kube-api-access-6td2m\") pod \"c167b57e-b623-4080-a770-d87233e00128\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.461533 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb\") pod \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\" (UID: \"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.461714 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-combined-ca-bundle\") pod \"c167b57e-b623-4080-a770-d87233e00128\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.461772 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-scripts\") pod \"c167b57e-b623-4080-a770-d87233e00128\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.461853 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c167b57e-b623-4080-a770-d87233e00128-logs\") pod \"c167b57e-b623-4080-a770-d87233e00128\" (UID: \"c167b57e-b623-4080-a770-d87233e00128\") " Dec 06 06:02:57 crc kubenswrapper[4957]: W1206 06:02:57.462083 4957 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e/volumes/kubernetes.io~configmap/ovsdbserver-sb Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.462100 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" (UID: "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.462240 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" (UID: "b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.462997 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c167b57e-b623-4080-a770-d87233e00128-logs" (OuterVolumeSpecName: "logs") pod "c167b57e-b623-4080-a770-d87233e00128" (UID: "c167b57e-b623-4080-a770-d87233e00128"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.463192 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c167b57e-b623-4080-a770-d87233e00128-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.463204 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.463212 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.463221 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.463382 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv2vv\" (UniqueName: \"kubernetes.io/projected/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-kube-api-access-hv2vv\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.463390 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.463398 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.466165 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c167b57e-b623-4080-a770-d87233e00128-kube-api-access-6td2m" (OuterVolumeSpecName: "kube-api-access-6td2m") pod "c167b57e-b623-4080-a770-d87233e00128" (UID: "c167b57e-b623-4080-a770-d87233e00128"). InnerVolumeSpecName "kube-api-access-6td2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.466442 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-scripts" (OuterVolumeSpecName: "scripts") pod "c167b57e-b623-4080-a770-d87233e00128" (UID: "c167b57e-b623-4080-a770-d87233e00128"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.490265 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c167b57e-b623-4080-a770-d87233e00128" (UID: "c167b57e-b623-4080-a770-d87233e00128"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.500900 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-config-data" (OuterVolumeSpecName: "config-data") pod "c167b57e-b623-4080-a770-d87233e00128" (UID: "c167b57e-b623-4080-a770-d87233e00128"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.565288 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.565322 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.565333 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c167b57e-b623-4080-a770-d87233e00128-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.565346 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6td2m\" (UniqueName: \"kubernetes.io/projected/c167b57e-b623-4080-a770-d87233e00128-kube-api-access-6td2m\") on node \"crc\" DevicePath \"\"" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.924116 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" event={"ID":"b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e","Type":"ContainerDied","Data":"1e23d53266c25f9c298b6f716d1ed2d06fb49b0c4528b1876ec27d586c3e5b93"} Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.924191 4957 scope.go:117] "RemoveContainer" containerID="0d35b8b3da1aab99f8dc72cf460458a1afe19bd0cdc7862664f55a885a4bbc56" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.924216 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-qmvb2" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.929055 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zsklh" event={"ID":"c167b57e-b623-4080-a770-d87233e00128","Type":"ContainerDied","Data":"349f856a84776999ffc0f70923b63abe332b236ccaf498cd6177459e9525d863"} Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.929111 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="349f856a84776999ffc0f70923b63abe332b236ccaf498cd6177459e9525d863" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.929080 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zsklh" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.968490 4957 scope.go:117] "RemoveContainer" containerID="f3866bd21b440a96b64e59c5a687f89e6f231e2bdbec125c33affd83c232ce3d" Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.968907 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qmvb2"] Dec 06 06:02:57 crc kubenswrapper[4957]: I1206 06:02:57.976883 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-qmvb2"] Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.061034 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-69dd5567fb-lm2jl"] Dec 06 06:02:58 crc kubenswrapper[4957]: E1206 06:02:58.061459 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerName="init" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.061471 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerName="init" Dec 06 06:02:58 crc kubenswrapper[4957]: E1206 06:02:58.061496 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerName="dnsmasq-dns" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.061502 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerName="dnsmasq-dns" Dec 06 06:02:58 crc kubenswrapper[4957]: E1206 06:02:58.061523 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c167b57e-b623-4080-a770-d87233e00128" containerName="placement-db-sync" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.061530 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c167b57e-b623-4080-a770-d87233e00128" containerName="placement-db-sync" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.061685 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" containerName="dnsmasq-dns" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.061694 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c167b57e-b623-4080-a770-d87233e00128" containerName="placement-db-sync" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.062603 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.064586 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.065312 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.065624 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.066669 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.066817 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-s85xq" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.079069 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69dd5567fb-lm2jl"] Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.184787 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-combined-ca-bundle\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.184859 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-public-tls-certs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.184895 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-config-data\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.184914 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqwnv\" (UniqueName: \"kubernetes.io/projected/8cada75b-b0d4-41c2-8d1f-3f146e28651d-kube-api-access-hqwnv\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.184935 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-scripts\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.184976 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-internal-tls-certs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.185027 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cada75b-b0d4-41c2-8d1f-3f146e28651d-logs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287213 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-internal-tls-certs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287354 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cada75b-b0d4-41c2-8d1f-3f146e28651d-logs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287439 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-combined-ca-bundle\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287496 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-public-tls-certs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287530 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-config-data\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287577 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqwnv\" (UniqueName: \"kubernetes.io/projected/8cada75b-b0d4-41c2-8d1f-3f146e28651d-kube-api-access-hqwnv\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287605 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-scripts\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.287704 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cada75b-b0d4-41c2-8d1f-3f146e28651d-logs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.292788 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-combined-ca-bundle\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.293163 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-scripts\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.293434 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-public-tls-certs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.311614 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-internal-tls-certs\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.311614 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cada75b-b0d4-41c2-8d1f-3f146e28651d-config-data\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.314298 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqwnv\" (UniqueName: \"kubernetes.io/projected/8cada75b-b0d4-41c2-8d1f-3f146e28651d-kube-api-access-hqwnv\") pod \"placement-69dd5567fb-lm2jl\" (UID: \"8cada75b-b0d4-41c2-8d1f-3f146e28651d\") " pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.385018 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.673540 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e" path="/var/lib/kubelet/pods/b5ddf589-9821-49a4-81c8-1fcc1d8e6d5e/volumes" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.792229 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-76c6cdbd84-nmkc9" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 06:02:58 crc kubenswrapper[4957]: I1206 06:02:58.869477 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69dd5567fb-lm2jl"] Dec 06 06:02:59 crc kubenswrapper[4957]: I1206 06:02:59.096229 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c76d956-bsr6l" podUID="cb06a82d-1610-4985-8168-c95f32ad2ff6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 06 06:02:59 crc kubenswrapper[4957]: I1206 06:02:59.950784 4957 generic.go:334] "Generic (PLEG): container finished" podID="7d3a3efb-65b9-4a90-bded-99969b537744" containerID="14b61a9a5561fe1188d263899b81f4fc8d0af35a3867fae10c46438c1746f01a" exitCode=0 Dec 06 06:02:59 crc kubenswrapper[4957]: I1206 06:02:59.950825 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f45dr" event={"ID":"7d3a3efb-65b9-4a90-bded-99969b537744","Type":"ContainerDied","Data":"14b61a9a5561fe1188d263899b81f4fc8d0af35a3867fae10c46438c1746f01a"} Dec 06 06:03:00 crc kubenswrapper[4957]: I1206 06:03:00.966564 4957 generic.go:334] "Generic (PLEG): container finished" podID="caa7934f-ce05-4cac-b198-b67ffa3e96ed" containerID="b70e91a0cf14b35a501421be0dec610a6bfe3f59df20f9c4f06da3e3c089f43d" exitCode=0 Dec 06 06:03:00 crc kubenswrapper[4957]: I1206 06:03:00.966664 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6szxq" event={"ID":"caa7934f-ce05-4cac-b198-b67ffa3e96ed","Type":"ContainerDied","Data":"b70e91a0cf14b35a501421be0dec610a6bfe3f59df20f9c4f06da3e3c089f43d"} Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.575624 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f45dr" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.578712 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hcvj\" (UniqueName: \"kubernetes.io/projected/7d3a3efb-65b9-4a90-bded-99969b537744-kube-api-access-4hcvj\") pod \"7d3a3efb-65b9-4a90-bded-99969b537744\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.578792 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-db-sync-config-data\") pod \"7d3a3efb-65b9-4a90-bded-99969b537744\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.578908 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-combined-ca-bundle\") pod \"7d3a3efb-65b9-4a90-bded-99969b537744\" (UID: \"7d3a3efb-65b9-4a90-bded-99969b537744\") " Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.589035 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7d3a3efb-65b9-4a90-bded-99969b537744" (UID: "7d3a3efb-65b9-4a90-bded-99969b537744"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.600182 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d3a3efb-65b9-4a90-bded-99969b537744-kube-api-access-4hcvj" (OuterVolumeSpecName: "kube-api-access-4hcvj") pod "7d3a3efb-65b9-4a90-bded-99969b537744" (UID: "7d3a3efb-65b9-4a90-bded-99969b537744"). InnerVolumeSpecName "kube-api-access-4hcvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.650145 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d3a3efb-65b9-4a90-bded-99969b537744" (UID: "7d3a3efb-65b9-4a90-bded-99969b537744"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.681218 4957 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.681258 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3a3efb-65b9-4a90-bded-99969b537744-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.681274 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hcvj\" (UniqueName: \"kubernetes.io/projected/7d3a3efb-65b9-4a90-bded-99969b537744-kube-api-access-4hcvj\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.994009 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f45dr" event={"ID":"7d3a3efb-65b9-4a90-bded-99969b537744","Type":"ContainerDied","Data":"490412cda4034d2908c83b64415ef0dbffaec5df78c79a63582560494e73612c"} Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.994046 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="490412cda4034d2908c83b64415ef0dbffaec5df78c79a63582560494e73612c" Dec 06 06:03:02 crc kubenswrapper[4957]: I1206 06:03:02.994114 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f45dr" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.845917 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5684df76ff-6tw89"] Dec 06 06:03:03 crc kubenswrapper[4957]: E1206 06:03:03.846474 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d3a3efb-65b9-4a90-bded-99969b537744" containerName="barbican-db-sync" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.846492 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d3a3efb-65b9-4a90-bded-99969b537744" containerName="barbican-db-sync" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.846762 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d3a3efb-65b9-4a90-bded-99969b537744" containerName="barbican-db-sync" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.848275 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.852019 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.852266 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.856721 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d4rs8" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.898992 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7bb9fc9b94-kknvp"] Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.900510 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.904920 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.912463 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5684df76ff-6tw89"] Dec 06 06:03:03 crc kubenswrapper[4957]: I1206 06:03:03.923895 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bb9fc9b94-kknvp"] Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009275 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1c9d4c1-173d-4db7-9986-2873c3280e2a-logs\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009609 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-logs\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009658 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-config-data-custom\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009692 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-combined-ca-bundle\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009716 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-config-data\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009765 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-combined-ca-bundle\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009800 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-config-data-custom\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009822 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5f9w\" (UniqueName: \"kubernetes.io/projected/b1c9d4c1-173d-4db7-9986-2873c3280e2a-kube-api-access-z5f9w\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009871 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-config-data\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.009904 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf95m\" (UniqueName: \"kubernetes.io/projected/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-kube-api-access-wf95m\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.019005 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69dd5567fb-lm2jl" event={"ID":"8cada75b-b0d4-41c2-8d1f-3f146e28651d","Type":"ContainerStarted","Data":"04f9fc6266d8cf9fca4c024a7ad4b99d2140c35bff0ef989687c85eb1d7b4f8b"} Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.039096 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6szxq" event={"ID":"caa7934f-ce05-4cac-b198-b67ffa3e96ed","Type":"ContainerDied","Data":"9a443aef8faa446220a003b2dddfb941436fb690ca411191f2a5d71f1df461fc"} Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.039154 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a443aef8faa446220a003b2dddfb941436fb690ca411191f2a5d71f1df461fc" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.047226 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6szxq" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.087894 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-7q9n2"] Dec 06 06:03:04 crc kubenswrapper[4957]: E1206 06:03:04.088483 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa7934f-ce05-4cac-b198-b67ffa3e96ed" containerName="cinder-db-sync" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.088525 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa7934f-ce05-4cac-b198-b67ffa3e96ed" containerName="cinder-db-sync" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.088858 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa7934f-ce05-4cac-b198-b67ffa3e96ed" containerName="cinder-db-sync" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.094254 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.116872 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-combined-ca-bundle\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.116917 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-config-data-custom\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.116940 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5f9w\" (UniqueName: \"kubernetes.io/projected/b1c9d4c1-173d-4db7-9986-2873c3280e2a-kube-api-access-z5f9w\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.116969 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-config-data\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.116995 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf95m\" (UniqueName: \"kubernetes.io/projected/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-kube-api-access-wf95m\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.117030 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1c9d4c1-173d-4db7-9986-2873c3280e2a-logs\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.117064 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-logs\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.117089 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-config-data-custom\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.117111 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-combined-ca-bundle\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.117127 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-config-data\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.118074 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1c9d4c1-173d-4db7-9986-2873c3280e2a-logs\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.119697 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-logs\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.134301 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-config-data\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.135115 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-combined-ca-bundle\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.136889 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-7q9n2"] Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.146070 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.146446 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.150405 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-combined-ca-bundle\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.151545 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-config-data\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.154343 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-config-data-custom\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.169611 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5f9w\" (UniqueName: \"kubernetes.io/projected/b1c9d4c1-173d-4db7-9986-2873c3280e2a-kube-api-access-z5f9w\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.169622 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1c9d4c1-173d-4db7-9986-2873c3280e2a-config-data-custom\") pod \"barbican-worker-5684df76ff-6tw89\" (UID: \"b1c9d4c1-173d-4db7-9986-2873c3280e2a\") " pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.172492 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf95m\" (UniqueName: \"kubernetes.io/projected/e3907b98-0bd2-4ad9-bb72-d1f651c16f68-kube-api-access-wf95m\") pod \"barbican-keystone-listener-7bb9fc9b94-kknvp\" (UID: \"e3907b98-0bd2-4ad9-bb72-d1f651c16f68\") " pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.181780 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5684df76ff-6tw89" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.223571 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-combined-ca-bundle\") pod \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.223623 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-config-data\") pod \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.223704 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-scripts\") pod \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.223738 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86chd\" (UniqueName: \"kubernetes.io/projected/caa7934f-ce05-4cac-b198-b67ffa3e96ed-kube-api-access-86chd\") pod \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.223763 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-db-sync-config-data\") pod \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.223908 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa7934f-ce05-4cac-b198-b67ffa3e96ed-etc-machine-id\") pod \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\" (UID: \"caa7934f-ce05-4cac-b198-b67ffa3e96ed\") " Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.224081 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.224133 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.224179 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n27gk\" (UniqueName: \"kubernetes.io/projected/35667d3c-590c-42cc-b1bd-6d14a0e94213-kube-api-access-n27gk\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.224235 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-config\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.224263 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.224291 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-svc\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.224583 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/caa7934f-ce05-4cac-b198-b67ffa3e96ed-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "caa7934f-ce05-4cac-b198-b67ffa3e96ed" (UID: "caa7934f-ce05-4cac-b198-b67ffa3e96ed"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.228311 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.239178 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.239656 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "caa7934f-ce05-4cac-b198-b67ffa3e96ed" (UID: "caa7934f-ce05-4cac-b198-b67ffa3e96ed"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.240141 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa7934f-ce05-4cac-b198-b67ffa3e96ed-kube-api-access-86chd" (OuterVolumeSpecName: "kube-api-access-86chd") pod "caa7934f-ce05-4cac-b198-b67ffa3e96ed" (UID: "caa7934f-ce05-4cac-b198-b67ffa3e96ed"). InnerVolumeSpecName "kube-api-access-86chd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.263957 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-scripts" (OuterVolumeSpecName: "scripts") pod "caa7934f-ce05-4cac-b198-b67ffa3e96ed" (UID: "caa7934f-ce05-4cac-b198-b67ffa3e96ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.269032 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c9cb5bbd6-kzmqd"] Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.282751 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.282867 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.298738 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c9cb5bbd6-kzmqd"] Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.308183 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caa7934f-ce05-4cac-b198-b67ffa3e96ed" (UID: "caa7934f-ce05-4cac-b198-b67ffa3e96ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.308425 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.325751 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-config\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.325805 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.326745 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-config\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329358 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-svc\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329397 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329546 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329697 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n27gk\" (UniqueName: \"kubernetes.io/projected/35667d3c-590c-42cc-b1bd-6d14a0e94213-kube-api-access-n27gk\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329695 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329862 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329876 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329885 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86chd\" (UniqueName: \"kubernetes.io/projected/caa7934f-ce05-4cac-b198-b67ffa3e96ed-kube-api-access-86chd\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329895 4957 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.329906 4957 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/caa7934f-ce05-4cac-b198-b67ffa3e96ed-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.331630 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.332186 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-svc\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.334446 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.385655 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n27gk\" (UniqueName: \"kubernetes.io/projected/35667d3c-590c-42cc-b1bd-6d14a0e94213-kube-api-access-n27gk\") pod \"dnsmasq-dns-85ff748b95-7q9n2\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.389054 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-config-data" (OuterVolumeSpecName: "config-data") pod "caa7934f-ce05-4cac-b198-b67ffa3e96ed" (UID: "caa7934f-ce05-4cac-b198-b67ffa3e96ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.431378 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.431981 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data-custom\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.432104 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-combined-ca-bundle\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.432312 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9159107-1504-4952-bf38-703c55a5a029-logs\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.432456 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvm9w\" (UniqueName: \"kubernetes.io/projected/a9159107-1504-4952-bf38-703c55a5a029-kube-api-access-tvm9w\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.432712 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa7934f-ce05-4cac-b198-b67ffa3e96ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.536227 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9159107-1504-4952-bf38-703c55a5a029-logs\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.536326 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvm9w\" (UniqueName: \"kubernetes.io/projected/a9159107-1504-4952-bf38-703c55a5a029-kube-api-access-tvm9w\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.536542 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.536563 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data-custom\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.536580 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-combined-ca-bundle\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.537887 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9159107-1504-4952-bf38-703c55a5a029-logs\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.539389 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.543553 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-combined-ca-bundle\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.555295 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data-custom\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.555300 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.570030 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvm9w\" (UniqueName: \"kubernetes.io/projected/a9159107-1504-4952-bf38-703c55a5a029-kube-api-access-tvm9w\") pod \"barbican-api-c9cb5bbd6-kzmqd\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: E1206 06:03:04.609327 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.626103 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.819444 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5684df76ff-6tw89"] Dec 06 06:03:04 crc kubenswrapper[4957]: W1206 06:03:04.821048 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1c9d4c1_173d_4db7_9986_2873c3280e2a.slice/crio-65e4b800ecef963bb1b9658dec0093cc13cb9a2cea14092ee8847fd11e8a6b0d WatchSource:0}: Error finding container 65e4b800ecef963bb1b9658dec0093cc13cb9a2cea14092ee8847fd11e8a6b0d: Status 404 returned error can't find the container with id 65e4b800ecef963bb1b9658dec0093cc13cb9a2cea14092ee8847fd11e8a6b0d Dec 06 06:03:04 crc kubenswrapper[4957]: I1206 06:03:04.908159 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bb9fc9b94-kknvp"] Dec 06 06:03:04 crc kubenswrapper[4957]: W1206 06:03:04.918490 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3907b98_0bd2_4ad9_bb72_d1f651c16f68.slice/crio-29bb069b4644c5f4b2811349c5b42506657eb021e563d1ee212c9c29ed495cdd WatchSource:0}: Error finding container 29bb069b4644c5f4b2811349c5b42506657eb021e563d1ee212c9c29ed495cdd: Status 404 returned error can't find the container with id 29bb069b4644c5f4b2811349c5b42506657eb021e563d1ee212c9c29ed495cdd Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.015654 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-7q9n2"] Dec 06 06:03:05 crc kubenswrapper[4957]: W1206 06:03:05.021907 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35667d3c_590c_42cc_b1bd_6d14a0e94213.slice/crio-fa0b3b292b58c182f28600aacf8835f5b3a096607b1f93917b190aec5cb13e64 WatchSource:0}: Error finding container fa0b3b292b58c182f28600aacf8835f5b3a096607b1f93917b190aec5cb13e64: Status 404 returned error can't find the container with id fa0b3b292b58c182f28600aacf8835f5b3a096607b1f93917b190aec5cb13e64 Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.048729 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5684df76ff-6tw89" event={"ID":"b1c9d4c1-173d-4db7-9986-2873c3280e2a","Type":"ContainerStarted","Data":"65e4b800ecef963bb1b9658dec0093cc13cb9a2cea14092ee8847fd11e8a6b0d"} Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.055318 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" event={"ID":"35667d3c-590c-42cc-b1bd-6d14a0e94213","Type":"ContainerStarted","Data":"fa0b3b292b58c182f28600aacf8835f5b3a096607b1f93917b190aec5cb13e64"} Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.061734 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerStarted","Data":"1c0b0f67a2d9949e4180115cf829725d72e84f720d59bf81dc837489d9ad5db3"} Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.061936 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="ceilometer-notification-agent" containerID="cri-o://11df35ebea1e26e949fe4e06981f15b1f30286868b24ce1b74fd58e1315a6fd6" gracePeriod=30 Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.062032 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.062345 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="proxy-httpd" containerID="cri-o://1c0b0f67a2d9949e4180115cf829725d72e84f720d59bf81dc837489d9ad5db3" gracePeriod=30 Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.062393 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="sg-core" containerID="cri-o://77234a9a9030cd0d0b92e72149b99781c3b5e68217509a6458ccbebbc9ff5c84" gracePeriod=30 Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.072116 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69dd5567fb-lm2jl" event={"ID":"8cada75b-b0d4-41c2-8d1f-3f146e28651d","Type":"ContainerStarted","Data":"e37d3d89bdf553ab7e74417666fb2c7e37100dee93e4b79af8d6e4c17f87016c"} Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.072357 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69dd5567fb-lm2jl" event={"ID":"8cada75b-b0d4-41c2-8d1f-3f146e28651d","Type":"ContainerStarted","Data":"8a53b63438e4a557a7509ab575278fea71055c08fe5133dfc66613a02a5e5012"} Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.073374 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.073504 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.090726 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" event={"ID":"e3907b98-0bd2-4ad9-bb72-d1f651c16f68","Type":"ContainerStarted","Data":"29bb069b4644c5f4b2811349c5b42506657eb021e563d1ee212c9c29ed495cdd"} Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.091360 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.091470 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.090793 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6szxq" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.221177 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c9cb5bbd6-kzmqd"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.230620 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-69dd5567fb-lm2jl" podStartSLOduration=7.230574496 podStartE2EDuration="7.230574496s" podCreationTimestamp="2025-12-06 06:02:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:05.138242139 +0000 UTC m=+1424.788509791" watchObservedRunningTime="2025-12-06 06:03:05.230574496 +0000 UTC m=+1424.880842128" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.297404 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.300272 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.302163 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.302420 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-h69mr" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.302562 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.306112 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.348014 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.362934 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-scripts\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.363005 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b7dv\" (UniqueName: \"kubernetes.io/projected/ee976330-e8f0-4e33-9374-bbce8c1064f6-kube-api-access-8b7dv\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.363045 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.363119 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.363164 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.363182 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee976330-e8f0-4e33-9374-bbce8c1064f6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.407816 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-7q9n2"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.445291 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9b95f"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.446852 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.465716 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-scripts\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.465762 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b7dv\" (UniqueName: \"kubernetes.io/projected/ee976330-e8f0-4e33-9374-bbce8c1064f6-kube-api-access-8b7dv\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.465793 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.465884 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.465930 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.465953 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee976330-e8f0-4e33-9374-bbce8c1064f6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.466022 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee976330-e8f0-4e33-9374-bbce8c1064f6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.466873 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9b95f"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.474660 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-scripts\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.474945 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.474992 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.478793 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.494955 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b7dv\" (UniqueName: \"kubernetes.io/projected/ee976330-e8f0-4e33-9374-bbce8c1064f6-kube-api-access-8b7dv\") pod \"cinder-scheduler-0\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.546641 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.548256 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.553139 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.561310 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.570003 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.570166 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.570413 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.570465 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-config\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.570531 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.570599 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch2sb\" (UniqueName: \"kubernetes.io/projected/f840f7f6-e5f8-4307-910f-5c30200ae1fb-kube-api-access-ch2sb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.654333 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.672687 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsvfl\" (UniqueName: \"kubernetes.io/projected/79a11f4c-0eee-488b-86bd-2b94d755bf0b-kube-api-access-wsvfl\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.672729 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-scripts\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.672775 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673685 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673725 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673764 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673793 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673824 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-config\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673887 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673937 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch2sb\" (UniqueName: \"kubernetes.io/projected/f840f7f6-e5f8-4307-910f-5c30200ae1fb-kube-api-access-ch2sb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.673980 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a11f4c-0eee-488b-86bd-2b94d755bf0b-logs\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.674006 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a11f4c-0eee-488b-86bd-2b94d755bf0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.674045 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.674075 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.674611 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.675246 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.675674 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-config\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.675777 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.691528 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch2sb\" (UniqueName: \"kubernetes.io/projected/f840f7f6-e5f8-4307-910f-5c30200ae1fb-kube-api-access-ch2sb\") pod \"dnsmasq-dns-5c9776ccc5-9b95f\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.776301 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.776349 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.776455 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a11f4c-0eee-488b-86bd-2b94d755bf0b-logs\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.776474 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a11f4c-0eee-488b-86bd-2b94d755bf0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.776506 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.776561 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsvfl\" (UniqueName: \"kubernetes.io/projected/79a11f4c-0eee-488b-86bd-2b94d755bf0b-kube-api-access-wsvfl\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.776580 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-scripts\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.777118 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a11f4c-0eee-488b-86bd-2b94d755bf0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.777342 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a11f4c-0eee-488b-86bd-2b94d755bf0b-logs\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.782356 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-scripts\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.782520 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.784949 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.786530 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.803968 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsvfl\" (UniqueName: \"kubernetes.io/projected/79a11f4c-0eee-488b-86bd-2b94d755bf0b-kube-api-access-wsvfl\") pod \"cinder-api-0\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " pod="openstack/cinder-api-0" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.838682 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:05 crc kubenswrapper[4957]: I1206 06:03:05.893487 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.156255 4957 generic.go:334] "Generic (PLEG): container finished" podID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerID="1c0b0f67a2d9949e4180115cf829725d72e84f720d59bf81dc837489d9ad5db3" exitCode=0 Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.156767 4957 generic.go:334] "Generic (PLEG): container finished" podID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerID="77234a9a9030cd0d0b92e72149b99781c3b5e68217509a6458ccbebbc9ff5c84" exitCode=2 Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.156847 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerDied","Data":"1c0b0f67a2d9949e4180115cf829725d72e84f720d59bf81dc837489d9ad5db3"} Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.156879 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerDied","Data":"77234a9a9030cd0d0b92e72149b99781c3b5e68217509a6458ccbebbc9ff5c84"} Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.171047 4957 generic.go:334] "Generic (PLEG): container finished" podID="35667d3c-590c-42cc-b1bd-6d14a0e94213" containerID="71ceced569706e9130e5222572c93d5c2560ba79de11cf5df80d21b4681eaa7a" exitCode=0 Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.171151 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" event={"ID":"35667d3c-590c-42cc-b1bd-6d14a0e94213","Type":"ContainerDied","Data":"71ceced569706e9130e5222572c93d5c2560ba79de11cf5df80d21b4681eaa7a"} Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.189268 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.204881 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" event={"ID":"a9159107-1504-4952-bf38-703c55a5a029","Type":"ContainerStarted","Data":"029139da834c4cb80541a8869fd993103c7108e8258fb58a7be99771eb2a5dbc"} Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.204947 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" event={"ID":"a9159107-1504-4952-bf38-703c55a5a029","Type":"ContainerStarted","Data":"fe09a4774559c1856570653c25fc450bcc3e78abc5d218f3d01920b10b355fe1"} Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.204958 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" event={"ID":"a9159107-1504-4952-bf38-703c55a5a029","Type":"ContainerStarted","Data":"6891e471f722e982c27ca8ada1660d09aea020a360b1470b915baf73e261fda0"} Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.205599 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.205785 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.252136 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" podStartSLOduration=2.25211081 podStartE2EDuration="2.25211081s" podCreationTimestamp="2025-12-06 06:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:06.244497265 +0000 UTC m=+1425.894764897" watchObservedRunningTime="2025-12-06 06:03:06.25211081 +0000 UTC m=+1425.902378452" Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.427367 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9b95f"] Dec 06 06:03:06 crc kubenswrapper[4957]: W1206 06:03:06.433017 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf840f7f6_e5f8_4307_910f_5c30200ae1fb.slice/crio-9b55a6be087026ac4dfb11a3ba631685c822870b9141e4dbce9e5472d5dfc944 WatchSource:0}: Error finding container 9b55a6be087026ac4dfb11a3ba631685c822870b9141e4dbce9e5472d5dfc944: Status 404 returned error can't find the container with id 9b55a6be087026ac4dfb11a3ba631685c822870b9141e4dbce9e5472d5dfc944 Dec 06 06:03:06 crc kubenswrapper[4957]: I1206 06:03:06.690408 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.218016 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee976330-e8f0-4e33-9374-bbce8c1064f6","Type":"ContainerStarted","Data":"81b63e435cb24271cd9bf2d21b2ab8cec588930d77430ba7baed1608c7a639a0"} Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.221524 4957 generic.go:334] "Generic (PLEG): container finished" podID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerID="9490628da4465835c0ddccb6293ac2d15881077d9cbc74f23b692569da9a0f93" exitCode=0 Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.221859 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" event={"ID":"f840f7f6-e5f8-4307-910f-5c30200ae1fb","Type":"ContainerDied","Data":"9490628da4465835c0ddccb6293ac2d15881077d9cbc74f23b692569da9a0f93"} Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.221926 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" event={"ID":"f840f7f6-e5f8-4307-910f-5c30200ae1fb","Type":"ContainerStarted","Data":"9b55a6be087026ac4dfb11a3ba631685c822870b9141e4dbce9e5472d5dfc944"} Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.470989 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.520808 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-sb\") pod \"35667d3c-590c-42cc-b1bd-6d14a0e94213\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.520879 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-nb\") pod \"35667d3c-590c-42cc-b1bd-6d14a0e94213\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.520928 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n27gk\" (UniqueName: \"kubernetes.io/projected/35667d3c-590c-42cc-b1bd-6d14a0e94213-kube-api-access-n27gk\") pod \"35667d3c-590c-42cc-b1bd-6d14a0e94213\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.521035 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-config\") pod \"35667d3c-590c-42cc-b1bd-6d14a0e94213\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.521079 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-svc\") pod \"35667d3c-590c-42cc-b1bd-6d14a0e94213\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.521152 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-swift-storage-0\") pod \"35667d3c-590c-42cc-b1bd-6d14a0e94213\" (UID: \"35667d3c-590c-42cc-b1bd-6d14a0e94213\") " Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.546445 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35667d3c-590c-42cc-b1bd-6d14a0e94213-kube-api-access-n27gk" (OuterVolumeSpecName: "kube-api-access-n27gk") pod "35667d3c-590c-42cc-b1bd-6d14a0e94213" (UID: "35667d3c-590c-42cc-b1bd-6d14a0e94213"). InnerVolumeSpecName "kube-api-access-n27gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.555125 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35667d3c-590c-42cc-b1bd-6d14a0e94213" (UID: "35667d3c-590c-42cc-b1bd-6d14a0e94213"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.558225 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-config" (OuterVolumeSpecName: "config") pod "35667d3c-590c-42cc-b1bd-6d14a0e94213" (UID: "35667d3c-590c-42cc-b1bd-6d14a0e94213"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.562665 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35667d3c-590c-42cc-b1bd-6d14a0e94213" (UID: "35667d3c-590c-42cc-b1bd-6d14a0e94213"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.564651 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35667d3c-590c-42cc-b1bd-6d14a0e94213" (UID: "35667d3c-590c-42cc-b1bd-6d14a0e94213"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.571587 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "35667d3c-590c-42cc-b1bd-6d14a0e94213" (UID: "35667d3c-590c-42cc-b1bd-6d14a0e94213"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.623721 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.623769 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.623785 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n27gk\" (UniqueName: \"kubernetes.io/projected/35667d3c-590c-42cc-b1bd-6d14a0e94213-kube-api-access-n27gk\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.623802 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.623819 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.623881 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35667d3c-590c-42cc-b1bd-6d14a0e94213-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.811754 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:07 crc kubenswrapper[4957]: I1206 06:03:07.811871 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.087245 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.249918 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" event={"ID":"f840f7f6-e5f8-4307-910f-5c30200ae1fb","Type":"ContainerStarted","Data":"8a02407f8552fa56458c1278085e94b00bc965d3cf46bbdb8cc4c13cde53b91f"} Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.250947 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.251912 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79a11f4c-0eee-488b-86bd-2b94d755bf0b","Type":"ContainerStarted","Data":"1f63df0215e508731056c35da420fb585381559a91f01ebb0b2ad04b34c8c47d"} Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.253608 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.262440 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-7q9n2" event={"ID":"35667d3c-590c-42cc-b1bd-6d14a0e94213","Type":"ContainerDied","Data":"fa0b3b292b58c182f28600aacf8835f5b3a096607b1f93917b190aec5cb13e64"} Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.262472 4957 scope.go:117] "RemoveContainer" containerID="71ceced569706e9130e5222572c93d5c2560ba79de11cf5df80d21b4681eaa7a" Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.278723 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" podStartSLOduration=3.278706874 podStartE2EDuration="3.278706874s" podCreationTimestamp="2025-12-06 06:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:08.276595967 +0000 UTC m=+1427.926863609" watchObservedRunningTime="2025-12-06 06:03:08.278706874 +0000 UTC m=+1427.928974506" Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.412932 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-7q9n2"] Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.421469 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-7q9n2"] Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.678642 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35667d3c-590c-42cc-b1bd-6d14a0e94213" path="/var/lib/kubelet/pods/35667d3c-590c-42cc-b1bd-6d14a0e94213/volumes" Dec 06 06:03:08 crc kubenswrapper[4957]: I1206 06:03:08.841314 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.274380 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79a11f4c-0eee-488b-86bd-2b94d755bf0b","Type":"ContainerStarted","Data":"2d3cb9b053cfa42db4595c8dcd9b5d72a5eef41bb343142ca0c24860babcad77"} Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.287408 4957 generic.go:334] "Generic (PLEG): container finished" podID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerID="11df35ebea1e26e949fe4e06981f15b1f30286868b24ce1b74fd58e1315a6fd6" exitCode=0 Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.287464 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerDied","Data":"11df35ebea1e26e949fe4e06981f15b1f30286868b24ce1b74fd58e1315a6fd6"} Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.304746 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" event={"ID":"e3907b98-0bd2-4ad9-bb72-d1f651c16f68","Type":"ContainerStarted","Data":"09d75a013841340f1aed96ca745401357e0da6e689b6004ee32558dd6a944c95"} Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.309341 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5684df76ff-6tw89" event={"ID":"b1c9d4c1-173d-4db7-9986-2873c3280e2a","Type":"ContainerStarted","Data":"eabdf5f11bb72dd3754bb0c68f68218390a84af0d798daaeeada32fd922ceb84"} Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.533784 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.658479 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-run-httpd\") pod \"3c3e4d9e-461d-468a-a9b2-108124de44ff\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.658563 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-scripts\") pod \"3c3e4d9e-461d-468a-a9b2-108124de44ff\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.658587 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-combined-ca-bundle\") pod \"3c3e4d9e-461d-468a-a9b2-108124de44ff\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.658684 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-config-data\") pod \"3c3e4d9e-461d-468a-a9b2-108124de44ff\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.658728 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-sg-core-conf-yaml\") pod \"3c3e4d9e-461d-468a-a9b2-108124de44ff\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.658761 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rvxs\" (UniqueName: \"kubernetes.io/projected/3c3e4d9e-461d-468a-a9b2-108124de44ff-kube-api-access-8rvxs\") pod \"3c3e4d9e-461d-468a-a9b2-108124de44ff\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.658800 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-log-httpd\") pod \"3c3e4d9e-461d-468a-a9b2-108124de44ff\" (UID: \"3c3e4d9e-461d-468a-a9b2-108124de44ff\") " Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.659069 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3c3e4d9e-461d-468a-a9b2-108124de44ff" (UID: "3c3e4d9e-461d-468a-a9b2-108124de44ff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.659245 4957 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.659311 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3c3e4d9e-461d-468a-a9b2-108124de44ff" (UID: "3c3e4d9e-461d-468a-a9b2-108124de44ff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.665228 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3e4d9e-461d-468a-a9b2-108124de44ff-kube-api-access-8rvxs" (OuterVolumeSpecName: "kube-api-access-8rvxs") pod "3c3e4d9e-461d-468a-a9b2-108124de44ff" (UID: "3c3e4d9e-461d-468a-a9b2-108124de44ff"). InnerVolumeSpecName "kube-api-access-8rvxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.669959 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-scripts" (OuterVolumeSpecName: "scripts") pod "3c3e4d9e-461d-468a-a9b2-108124de44ff" (UID: "3c3e4d9e-461d-468a-a9b2-108124de44ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.700175 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3c3e4d9e-461d-468a-a9b2-108124de44ff" (UID: "3c3e4d9e-461d-468a-a9b2-108124de44ff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.733163 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c3e4d9e-461d-468a-a9b2-108124de44ff" (UID: "3c3e4d9e-461d-468a-a9b2-108124de44ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.760607 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.760636 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.760646 4957 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.760682 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rvxs\" (UniqueName: \"kubernetes.io/projected/3c3e4d9e-461d-468a-a9b2-108124de44ff-kube-api-access-8rvxs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.760693 4957 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c3e4d9e-461d-468a-a9b2-108124de44ff-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.764607 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-config-data" (OuterVolumeSpecName: "config-data") pod "3c3e4d9e-461d-468a-a9b2-108124de44ff" (UID: "3c3e4d9e-461d-468a-a9b2-108124de44ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:09 crc kubenswrapper[4957]: I1206 06:03:09.868141 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c3e4d9e-461d-468a-a9b2-108124de44ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.327293 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c3e4d9e-461d-468a-a9b2-108124de44ff","Type":"ContainerDied","Data":"956088ec133a52d47d84d71b74a7bcaf2a43c1cd18b6bef443aac1f95f2889f8"} Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.327629 4957 scope.go:117] "RemoveContainer" containerID="1c0b0f67a2d9949e4180115cf829725d72e84f720d59bf81dc837489d9ad5db3" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.327327 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.340623 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" event={"ID":"e3907b98-0bd2-4ad9-bb72-d1f651c16f68","Type":"ContainerStarted","Data":"6981a6421906d422d16722ef540543f4bf4d54f13c30512c385015e4be28fbdd"} Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.358451 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5684df76ff-6tw89" event={"ID":"b1c9d4c1-173d-4db7-9986-2873c3280e2a","Type":"ContainerStarted","Data":"23b96afe1740ee21b46ac3ccd61783190a5378aa3bfcaa3381946b38519839b1"} Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.372211 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7bb9fc9b94-kknvp" podStartSLOduration=4.831823984 podStartE2EDuration="7.372184987s" podCreationTimestamp="2025-12-06 06:03:03 +0000 UTC" firstStartedPulling="2025-12-06 06:03:04.921652325 +0000 UTC m=+1424.571919957" lastFinishedPulling="2025-12-06 06:03:07.462013328 +0000 UTC m=+1427.112280960" observedRunningTime="2025-12-06 06:03:10.356971335 +0000 UTC m=+1430.007238987" watchObservedRunningTime="2025-12-06 06:03:10.372184987 +0000 UTC m=+1430.022452619" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.412032 4957 scope.go:117] "RemoveContainer" containerID="77234a9a9030cd0d0b92e72149b99781c3b5e68217509a6458ccbebbc9ff5c84" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.470333 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api-log" containerID="cri-o://2d3cb9b053cfa42db4595c8dcd9b5d72a5eef41bb343142ca0c24860babcad77" gracePeriod=30 Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.470407 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api" containerID="cri-o://fcbb924b48bab605169fb7c1ffce71edb02b34bc9da5fcad9f714e8961c8e6ef" gracePeriod=30 Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.470373 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79a11f4c-0eee-488b-86bd-2b94d755bf0b","Type":"ContainerStarted","Data":"fcbb924b48bab605169fb7c1ffce71edb02b34bc9da5fcad9f714e8961c8e6ef"} Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.470471 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.487707 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee976330-e8f0-4e33-9374-bbce8c1064f6","Type":"ContainerStarted","Data":"d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd"} Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.502583 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.534170 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.550903 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:10 crc kubenswrapper[4957]: E1206 06:03:10.551400 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="sg-core" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551421 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="sg-core" Dec 06 06:03:10 crc kubenswrapper[4957]: E1206 06:03:10.551456 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35667d3c-590c-42cc-b1bd-6d14a0e94213" containerName="init" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551464 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="35667d3c-590c-42cc-b1bd-6d14a0e94213" containerName="init" Dec 06 06:03:10 crc kubenswrapper[4957]: E1206 06:03:10.551495 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="ceilometer-notification-agent" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551503 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="ceilometer-notification-agent" Dec 06 06:03:10 crc kubenswrapper[4957]: E1206 06:03:10.551519 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="proxy-httpd" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551526 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="proxy-httpd" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551725 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="35667d3c-590c-42cc-b1bd-6d14a0e94213" containerName="init" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551757 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="sg-core" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551772 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="proxy-httpd" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.551791 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" containerName="ceilometer-notification-agent" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.553803 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.556293 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5684df76ff-6tw89" podStartSLOduration=4.947558131 podStartE2EDuration="7.556278025s" podCreationTimestamp="2025-12-06 06:03:03 +0000 UTC" firstStartedPulling="2025-12-06 06:03:04.823256095 +0000 UTC m=+1424.473523727" lastFinishedPulling="2025-12-06 06:03:07.431975989 +0000 UTC m=+1427.082243621" observedRunningTime="2025-12-06 06:03:10.496588091 +0000 UTC m=+1430.146855753" watchObservedRunningTime="2025-12-06 06:03:10.556278025 +0000 UTC m=+1430.206545657" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.556789 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.557336 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.564027 4957 scope.go:117] "RemoveContainer" containerID="11df35ebea1e26e949fe4e06981f15b1f30286868b24ce1b74fd58e1315a6fd6" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.568665 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.572439 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.572421032 podStartE2EDuration="5.572421032s" podCreationTimestamp="2025-12-06 06:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:10.533861669 +0000 UTC m=+1430.184129311" watchObservedRunningTime="2025-12-06 06:03:10.572421032 +0000 UTC m=+1430.222688674" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.685373 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3e4d9e-461d-468a-a9b2-108124de44ff" path="/var/lib/kubelet/pods/3c3e4d9e-461d-468a-a9b2-108124de44ff/volumes" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.700810 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.700856 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-log-httpd\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.702561 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn67g\" (UniqueName: \"kubernetes.io/projected/b38c57a0-87fd-41ad-9108-98cc86622c56-kube-api-access-fn67g\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.702663 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.702792 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-config-data\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.702908 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-run-httpd\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.703048 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-scripts\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.804775 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn67g\" (UniqueName: \"kubernetes.io/projected/b38c57a0-87fd-41ad-9108-98cc86622c56-kube-api-access-fn67g\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.805174 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.805222 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-config-data\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.805255 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-run-httpd\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.805302 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-scripts\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.805365 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.805381 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-log-httpd\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.805816 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-log-httpd\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.810202 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-run-httpd\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.815708 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.820590 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-scripts\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.831140 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.831792 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-config-data\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.834679 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn67g\" (UniqueName: \"kubernetes.io/projected/b38c57a0-87fd-41ad-9108-98cc86622c56-kube-api-access-fn67g\") pod \"ceilometer-0\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " pod="openstack/ceilometer-0" Dec 06 06:03:10 crc kubenswrapper[4957]: I1206 06:03:10.906897 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.321938 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.358343 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c9f9fb958-p7bk5"] Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.359901 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.375178 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.375326 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c9f9fb958-p7bk5"] Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.375697 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.424821 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.504078 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerStarted","Data":"49b54e089892aeb4c9b75c601621bc3beb3d4de162e49b28fb367fc0edec8252"} Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.505670 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79a11f4c-0eee-488b-86bd-2b94d755bf0b","Type":"ContainerDied","Data":"fcbb924b48bab605169fb7c1ffce71edb02b34bc9da5fcad9f714e8961c8e6ef"} Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.506249 4957 generic.go:334] "Generic (PLEG): container finished" podID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerID="fcbb924b48bab605169fb7c1ffce71edb02b34bc9da5fcad9f714e8961c8e6ef" exitCode=0 Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.506287 4957 generic.go:334] "Generic (PLEG): container finished" podID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerID="2d3cb9b053cfa42db4595c8dcd9b5d72a5eef41bb343142ca0c24860babcad77" exitCode=143 Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.506352 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79a11f4c-0eee-488b-86bd-2b94d755bf0b","Type":"ContainerDied","Data":"2d3cb9b053cfa42db4595c8dcd9b5d72a5eef41bb343142ca0c24860babcad77"} Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.508490 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee976330-e8f0-4e33-9374-bbce8c1064f6","Type":"ContainerStarted","Data":"60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87"} Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.528486 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-combined-ca-bundle\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.528575 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4w9m\" (UniqueName: \"kubernetes.io/projected/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-kube-api-access-n4w9m\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.528611 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-internal-tls-certs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.528638 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-public-tls-certs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.528682 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-config-data-custom\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.528713 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-logs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.528768 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-config-data\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.537969 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.82786984 podStartE2EDuration="6.537953344s" podCreationTimestamp="2025-12-06 06:03:05 +0000 UTC" firstStartedPulling="2025-12-06 06:03:06.222489982 +0000 UTC m=+1425.872757614" lastFinishedPulling="2025-12-06 06:03:08.932573496 +0000 UTC m=+1428.582841118" observedRunningTime="2025-12-06 06:03:11.526771011 +0000 UTC m=+1431.177038643" watchObservedRunningTime="2025-12-06 06:03:11.537953344 +0000 UTC m=+1431.188220966" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.610549 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.631389 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-combined-ca-bundle\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.633205 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4w9m\" (UniqueName: \"kubernetes.io/projected/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-kube-api-access-n4w9m\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.634059 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-internal-tls-certs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.634662 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-public-tls-certs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.634774 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-config-data-custom\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.635280 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-logs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.635377 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-config-data\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.636376 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-logs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.641822 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-public-tls-certs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.645057 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-config-data\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.646714 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-internal-tls-certs\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.657334 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-config-data-custom\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.666516 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-combined-ca-bundle\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.693561 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4w9m\" (UniqueName: \"kubernetes.io/projected/59b9c817-9c5c-4f08-9ca9-9d54e1d0912b-kube-api-access-n4w9m\") pod \"barbican-api-c9f9fb958-p7bk5\" (UID: \"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b\") " pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.713319 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:11 crc kubenswrapper[4957]: I1206 06:03:11.953388 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.052627 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a11f4c-0eee-488b-86bd-2b94d755bf0b-etc-machine-id\") pod \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.052670 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsvfl\" (UniqueName: \"kubernetes.io/projected/79a11f4c-0eee-488b-86bd-2b94d755bf0b-kube-api-access-wsvfl\") pod \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.052701 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-scripts\") pod \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.052726 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data-custom\") pod \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.052794 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a11f4c-0eee-488b-86bd-2b94d755bf0b-logs\") pod \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.052876 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-combined-ca-bundle\") pod \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.052892 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data\") pod \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\" (UID: \"79a11f4c-0eee-488b-86bd-2b94d755bf0b\") " Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.053286 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79a11f4c-0eee-488b-86bd-2b94d755bf0b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "79a11f4c-0eee-488b-86bd-2b94d755bf0b" (UID: "79a11f4c-0eee-488b-86bd-2b94d755bf0b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.054688 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79a11f4c-0eee-488b-86bd-2b94d755bf0b-logs" (OuterVolumeSpecName: "logs") pod "79a11f4c-0eee-488b-86bd-2b94d755bf0b" (UID: "79a11f4c-0eee-488b-86bd-2b94d755bf0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.065985 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a11f4c-0eee-488b-86bd-2b94d755bf0b-kube-api-access-wsvfl" (OuterVolumeSpecName: "kube-api-access-wsvfl") pod "79a11f4c-0eee-488b-86bd-2b94d755bf0b" (UID: "79a11f4c-0eee-488b-86bd-2b94d755bf0b"). InnerVolumeSpecName "kube-api-access-wsvfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.090994 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "79a11f4c-0eee-488b-86bd-2b94d755bf0b" (UID: "79a11f4c-0eee-488b-86bd-2b94d755bf0b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.091055 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-scripts" (OuterVolumeSpecName: "scripts") pod "79a11f4c-0eee-488b-86bd-2b94d755bf0b" (UID: "79a11f4c-0eee-488b-86bd-2b94d755bf0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.129736 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data" (OuterVolumeSpecName: "config-data") pod "79a11f4c-0eee-488b-86bd-2b94d755bf0b" (UID: "79a11f4c-0eee-488b-86bd-2b94d755bf0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.156196 4957 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a11f4c-0eee-488b-86bd-2b94d755bf0b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.156236 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsvfl\" (UniqueName: \"kubernetes.io/projected/79a11f4c-0eee-488b-86bd-2b94d755bf0b-kube-api-access-wsvfl\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.156251 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.156265 4957 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.156275 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a11f4c-0eee-488b-86bd-2b94d755bf0b-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.156286 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.179198 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79a11f4c-0eee-488b-86bd-2b94d755bf0b" (UID: "79a11f4c-0eee-488b-86bd-2b94d755bf0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.202022 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.258436 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a11f4c-0eee-488b-86bd-2b94d755bf0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.298548 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c9f9fb958-p7bk5"] Dec 06 06:03:12 crc kubenswrapper[4957]: W1206 06:03:12.329282 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59b9c817_9c5c_4f08_9ca9_9d54e1d0912b.slice/crio-df7fb2206a5755e089c2f621bbac88b03b008390cfd79c9bca1111173119f010 WatchSource:0}: Error finding container df7fb2206a5755e089c2f621bbac88b03b008390cfd79c9bca1111173119f010: Status 404 returned error can't find the container with id df7fb2206a5755e089c2f621bbac88b03b008390cfd79c9bca1111173119f010 Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.546008 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerStarted","Data":"f9ea6af81eeac4d2567ee02d47f19d6b17a9911b3df146d3d748f1c89739bcf7"} Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.549050 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9f9fb958-p7bk5" event={"ID":"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b","Type":"ContainerStarted","Data":"eb341a2ae1bd3b62011a41d17cc889603ed8fa403f3ad5795bbf9f0becc5eb1e"} Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.549081 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9f9fb958-p7bk5" event={"ID":"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b","Type":"ContainerStarted","Data":"df7fb2206a5755e089c2f621bbac88b03b008390cfd79c9bca1111173119f010"} Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.562976 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"79a11f4c-0eee-488b-86bd-2b94d755bf0b","Type":"ContainerDied","Data":"1f63df0215e508731056c35da420fb585381559a91f01ebb0b2ad04b34c8c47d"} Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.563050 4957 scope.go:117] "RemoveContainer" containerID="fcbb924b48bab605169fb7c1ffce71edb02b34bc9da5fcad9f714e8961c8e6ef" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.563303 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.939684 4957 scope.go:117] "RemoveContainer" containerID="2d3cb9b053cfa42db4595c8dcd9b5d72a5eef41bb343142ca0c24860babcad77" Dec 06 06:03:12 crc kubenswrapper[4957]: I1206 06:03:12.966195 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:03:13 crc kubenswrapper[4957]: E1206 06:03:13.025433 4957 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod430ef178_b9c0_4759_979b_92620ceca29f.slice/crio-df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90caf120_13de_4d6f_9e95_2954a448e874.slice/crio-a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79a11f4c_0eee_488b_86bd_2b94d755bf0b.slice/crio-1f63df0215e508731056c35da420fb585381559a91f01ebb0b2ad04b34c8c47d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod430ef178_b9c0_4759_979b_92620ceca29f.slice/crio-conmon-df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.035903 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.047525 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.060546 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:13 crc kubenswrapper[4957]: E1206 06:03:13.061432 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061451 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon" Dec 06 06:03:13 crc kubenswrapper[4957]: E1206 06:03:13.061479 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon-log" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061487 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon-log" Dec 06 06:03:13 crc kubenswrapper[4957]: E1206 06:03:13.061498 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api-log" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061506 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api-log" Dec 06 06:03:13 crc kubenswrapper[4957]: E1206 06:03:13.061534 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061541 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061724 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061737 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon-log" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061749 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="90caf120-13de-4d6f-9e95-2954a448e874" containerName="horizon" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.061767 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" containerName="cinder-api-log" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.063872 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.066055 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.066265 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.066730 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.091505 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.110342 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-scripts\") pod \"90caf120-13de-4d6f-9e95-2954a448e874\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.110434 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90caf120-13de-4d6f-9e95-2954a448e874-horizon-secret-key\") pod \"90caf120-13de-4d6f-9e95-2954a448e874\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.110544 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-config-data\") pod \"90caf120-13de-4d6f-9e95-2954a448e874\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.110568 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90caf120-13de-4d6f-9e95-2954a448e874-logs\") pod \"90caf120-13de-4d6f-9e95-2954a448e874\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.110595 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw5tn\" (UniqueName: \"kubernetes.io/projected/90caf120-13de-4d6f-9e95-2954a448e874-kube-api-access-dw5tn\") pod \"90caf120-13de-4d6f-9e95-2954a448e874\" (UID: \"90caf120-13de-4d6f-9e95-2954a448e874\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.122028 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90caf120-13de-4d6f-9e95-2954a448e874-logs" (OuterVolumeSpecName: "logs") pod "90caf120-13de-4d6f-9e95-2954a448e874" (UID: "90caf120-13de-4d6f-9e95-2954a448e874"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.126922 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90caf120-13de-4d6f-9e95-2954a448e874-kube-api-access-dw5tn" (OuterVolumeSpecName: "kube-api-access-dw5tn") pod "90caf120-13de-4d6f-9e95-2954a448e874" (UID: "90caf120-13de-4d6f-9e95-2954a448e874"). InnerVolumeSpecName "kube-api-access-dw5tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.135103 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90caf120-13de-4d6f-9e95-2954a448e874-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "90caf120-13de-4d6f-9e95-2954a448e874" (UID: "90caf120-13de-4d6f-9e95-2954a448e874"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.151206 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-scripts" (OuterVolumeSpecName: "scripts") pod "90caf120-13de-4d6f-9e95-2954a448e874" (UID: "90caf120-13de-4d6f-9e95-2954a448e874"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.161863 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-config-data" (OuterVolumeSpecName: "config-data") pod "90caf120-13de-4d6f-9e95-2954a448e874" (UID: "90caf120-13de-4d6f-9e95-2954a448e874"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.221005 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.221082 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.221136 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxt29\" (UniqueName: \"kubernetes.io/projected/c21557b2-c3ad-497a-ad42-ca81fff434bf-kube-api-access-cxt29\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.221153 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.221171 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21557b2-c3ad-497a-ad42-ca81fff434bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.221193 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c21557b2-c3ad-497a-ad42-ca81fff434bf-logs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.221344 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.223143 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-scripts\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.223217 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-config-data\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.223376 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.223398 4957 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90caf120-13de-4d6f-9e95-2954a448e874-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.223414 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90caf120-13de-4d6f-9e95-2954a448e874-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.223425 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90caf120-13de-4d6f-9e95-2954a448e874-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.223438 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw5tn\" (UniqueName: \"kubernetes.io/projected/90caf120-13de-4d6f-9e95-2954a448e874-kube-api-access-dw5tn\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.248147 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326513 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-scripts\") pod \"430ef178-b9c0-4759-979b-92620ceca29f\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326570 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-config-data\") pod \"430ef178-b9c0-4759-979b-92620ceca29f\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326608 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430ef178-b9c0-4759-979b-92620ceca29f-logs\") pod \"430ef178-b9c0-4759-979b-92620ceca29f\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326628 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/430ef178-b9c0-4759-979b-92620ceca29f-horizon-secret-key\") pod \"430ef178-b9c0-4759-979b-92620ceca29f\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326658 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7rsp\" (UniqueName: \"kubernetes.io/projected/430ef178-b9c0-4759-979b-92620ceca29f-kube-api-access-m7rsp\") pod \"430ef178-b9c0-4759-979b-92620ceca29f\" (UID: \"430ef178-b9c0-4759-979b-92620ceca29f\") " Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326926 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxt29\" (UniqueName: \"kubernetes.io/projected/c21557b2-c3ad-497a-ad42-ca81fff434bf-kube-api-access-cxt29\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326958 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.326981 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21557b2-c3ad-497a-ad42-ca81fff434bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.327011 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c21557b2-c3ad-497a-ad42-ca81fff434bf-logs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.327047 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.327093 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-scripts\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.327119 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-config-data\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.327174 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.327227 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.328163 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/430ef178-b9c0-4759-979b-92620ceca29f-logs" (OuterVolumeSpecName: "logs") pod "430ef178-b9c0-4759-979b-92620ceca29f" (UID: "430ef178-b9c0-4759-979b-92620ceca29f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.331050 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21557b2-c3ad-497a-ad42-ca81fff434bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.332154 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c21557b2-c3ad-497a-ad42-ca81fff434bf-logs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.332926 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.334562 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.337531 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430ef178-b9c0-4759-979b-92620ceca29f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "430ef178-b9c0-4759-979b-92620ceca29f" (UID: "430ef178-b9c0-4759-979b-92620ceca29f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.337993 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-config-data\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.342172 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-scripts\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.342325 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.344932 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21557b2-c3ad-497a-ad42-ca81fff434bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.348001 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/430ef178-b9c0-4759-979b-92620ceca29f-kube-api-access-m7rsp" (OuterVolumeSpecName: "kube-api-access-m7rsp") pod "430ef178-b9c0-4759-979b-92620ceca29f" (UID: "430ef178-b9c0-4759-979b-92620ceca29f"). InnerVolumeSpecName "kube-api-access-m7rsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.352020 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxt29\" (UniqueName: \"kubernetes.io/projected/c21557b2-c3ad-497a-ad42-ca81fff434bf-kube-api-access-cxt29\") pod \"cinder-api-0\" (UID: \"c21557b2-c3ad-497a-ad42-ca81fff434bf\") " pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.358776 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-scripts" (OuterVolumeSpecName: "scripts") pod "430ef178-b9c0-4759-979b-92620ceca29f" (UID: "430ef178-b9c0-4759-979b-92620ceca29f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.366958 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-config-data" (OuterVolumeSpecName: "config-data") pod "430ef178-b9c0-4759-979b-92620ceca29f" (UID: "430ef178-b9c0-4759-979b-92620ceca29f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.437435 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.437497 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/430ef178-b9c0-4759-979b-92620ceca29f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.437518 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430ef178-b9c0-4759-979b-92620ceca29f-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.437536 4957 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/430ef178-b9c0-4759-979b-92620ceca29f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.437557 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7rsp\" (UniqueName: \"kubernetes.io/projected/430ef178-b9c0-4759-979b-92620ceca29f-kube-api-access-m7rsp\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.546284 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.573998 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerStarted","Data":"75a254a9e897fcbd2463c3b029ba7056131f1bd241f50428f75ca10c32031549"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.578046 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9f9fb958-p7bk5" event={"ID":"59b9c817-9c5c-4f08-9ca9-9d54e1d0912b","Type":"ContainerStarted","Data":"e457958dd6a2f7520e57af6d9fb4acff73d6cb855c1b75a8f88b37e070127e0b"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.579712 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.579752 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.581339 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.582712 4957 generic.go:334] "Generic (PLEG): container finished" podID="430ef178-b9c0-4759-979b-92620ceca29f" containerID="2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0" exitCode=137 Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.582749 4957 generic.go:334] "Generic (PLEG): container finished" podID="430ef178-b9c0-4759-979b-92620ceca29f" containerID="df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933" exitCode=137 Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.582765 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bcdd755cc-4mppd" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.582825 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd755cc-4mppd" event={"ID":"430ef178-b9c0-4759-979b-92620ceca29f","Type":"ContainerDied","Data":"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.582908 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd755cc-4mppd" event={"ID":"430ef178-b9c0-4759-979b-92620ceca29f","Type":"ContainerDied","Data":"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.582929 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bcdd755cc-4mppd" event={"ID":"430ef178-b9c0-4759-979b-92620ceca29f","Type":"ContainerDied","Data":"cb7aadf5ba55384de2fa9075f7b8fc207c36048ec9832d89a3aff7924021d5ef"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.582953 4957 scope.go:117] "RemoveContainer" containerID="2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.600020 4957 generic.go:334] "Generic (PLEG): container finished" podID="90caf120-13de-4d6f-9e95-2954a448e874" containerID="a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff" exitCode=137 Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.600066 4957 generic.go:334] "Generic (PLEG): container finished" podID="90caf120-13de-4d6f-9e95-2954a448e874" containerID="fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658" exitCode=137 Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.600099 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54c4596cfc-p2m5g" event={"ID":"90caf120-13de-4d6f-9e95-2954a448e874","Type":"ContainerDied","Data":"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.600140 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54c4596cfc-p2m5g" event={"ID":"90caf120-13de-4d6f-9e95-2954a448e874","Type":"ContainerDied","Data":"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.600162 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54c4596cfc-p2m5g" event={"ID":"90caf120-13de-4d6f-9e95-2954a448e874","Type":"ContainerDied","Data":"3ffaf03690cdea7f1f6824feaad6e3521e68c06166606e8339b02c26e99a5f82"} Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.600249 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54c4596cfc-p2m5g" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.612180 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c9f9fb958-p7bk5" podStartSLOduration=2.612161378 podStartE2EDuration="2.612161378s" podCreationTimestamp="2025-12-06 06:03:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:13.607890723 +0000 UTC m=+1433.258158385" watchObservedRunningTime="2025-12-06 06:03:13.612161378 +0000 UTC m=+1433.262429010" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.679895 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54c4596cfc-p2m5g"] Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.693899 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-54c4596cfc-p2m5g"] Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.717898 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bcdd755cc-4mppd"] Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.719654 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-bcdd755cc-4mppd"] Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.871801 4957 scope.go:117] "RemoveContainer" containerID="df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.892407 4957 scope.go:117] "RemoveContainer" containerID="2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0" Dec 06 06:03:13 crc kubenswrapper[4957]: E1206 06:03:13.892918 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0\": container with ID starting with 2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0 not found: ID does not exist" containerID="2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.892985 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0"} err="failed to get container status \"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0\": rpc error: code = NotFound desc = could not find container \"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0\": container with ID starting with 2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0 not found: ID does not exist" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.893019 4957 scope.go:117] "RemoveContainer" containerID="df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933" Dec 06 06:03:13 crc kubenswrapper[4957]: E1206 06:03:13.893325 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933\": container with ID starting with df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933 not found: ID does not exist" containerID="df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.893377 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933"} err="failed to get container status \"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933\": rpc error: code = NotFound desc = could not find container \"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933\": container with ID starting with df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933 not found: ID does not exist" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.893406 4957 scope.go:117] "RemoveContainer" containerID="2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.893680 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0"} err="failed to get container status \"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0\": rpc error: code = NotFound desc = could not find container \"2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0\": container with ID starting with 2023e369f50a59c00029530f4b700a5ecdc903bad29c0365b584337a19cb21a0 not found: ID does not exist" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.893709 4957 scope.go:117] "RemoveContainer" containerID="df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.893977 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933"} err="failed to get container status \"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933\": rpc error: code = NotFound desc = could not find container \"df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933\": container with ID starting with df284274b1441cac689d24aef8d71ad8ebc278ec943f33c0d0441ce0838ea933 not found: ID does not exist" Dec 06 06:03:13 crc kubenswrapper[4957]: I1206 06:03:13.894002 4957 scope.go:117] "RemoveContainer" containerID="a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.078758 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.087731 4957 scope.go:117] "RemoveContainer" containerID="fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.239526 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5c76d956-bsr6l" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.300153 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76c6cdbd84-nmkc9"] Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.307900 4957 scope.go:117] "RemoveContainer" containerID="a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff" Dec 06 06:03:14 crc kubenswrapper[4957]: E1206 06:03:14.308542 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff\": container with ID starting with a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff not found: ID does not exist" containerID="a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.308593 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff"} err="failed to get container status \"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff\": rpc error: code = NotFound desc = could not find container \"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff\": container with ID starting with a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff not found: ID does not exist" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.308614 4957 scope.go:117] "RemoveContainer" containerID="fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658" Dec 06 06:03:14 crc kubenswrapper[4957]: E1206 06:03:14.309025 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658\": container with ID starting with fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658 not found: ID does not exist" containerID="fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.309074 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658"} err="failed to get container status \"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658\": rpc error: code = NotFound desc = could not find container \"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658\": container with ID starting with fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658 not found: ID does not exist" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.309088 4957 scope.go:117] "RemoveContainer" containerID="a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.309328 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff"} err="failed to get container status \"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff\": rpc error: code = NotFound desc = could not find container \"a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff\": container with ID starting with a4192c53a3a47a5e999b05794ed5034355df0366ed1aeed9021cef00013ab1ff not found: ID does not exist" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.309346 4957 scope.go:117] "RemoveContainer" containerID="fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.309636 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658"} err="failed to get container status \"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658\": rpc error: code = NotFound desc = could not find container \"fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658\": container with ID starting with fc05d1c6276ae46c88df9848a95ef58a400e75bf76e33ee58cb87f59d7ba3658 not found: ID does not exist" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.379351 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-575d75c97c-wblw4" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.449052 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-547bff997b-g5msk"] Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.450177 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-547bff997b-g5msk" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-api" containerID="cri-o://5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00" gracePeriod=30 Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.450608 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-547bff997b-g5msk" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-httpd" containerID="cri-o://9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f" gracePeriod=30 Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.625085 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerStarted","Data":"c67ec980d4e6c56f4805e1d461b1e25fdf1bbe6c8597c01c9000e4d5a83bf411"} Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.630819 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c21557b2-c3ad-497a-ad42-ca81fff434bf","Type":"ContainerStarted","Data":"f0139fccfb9195b2a2dc1a3c2ac42b9e40835f74ed5727ab91f14e8ece0b4c37"} Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.631035 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76c6cdbd84-nmkc9" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon-log" containerID="cri-o://1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48" gracePeriod=30 Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.631208 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76c6cdbd84-nmkc9" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" containerID="cri-o://da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b" gracePeriod=30 Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.676392 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="430ef178-b9c0-4759-979b-92620ceca29f" path="/var/lib/kubelet/pods/430ef178-b9c0-4759-979b-92620ceca29f/volumes" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.677247 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a11f4c-0eee-488b-86bd-2b94d755bf0b" path="/var/lib/kubelet/pods/79a11f4c-0eee-488b-86bd-2b94d755bf0b/volumes" Dec 06 06:03:14 crc kubenswrapper[4957]: I1206 06:03:14.678380 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90caf120-13de-4d6f-9e95-2954a448e874" path="/var/lib/kubelet/pods/90caf120-13de-4d6f-9e95-2954a448e874/volumes" Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.644152 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c21557b2-c3ad-497a-ad42-ca81fff434bf","Type":"ContainerStarted","Data":"d79a489bddb522c65b156d14ab7fbd9cc33b4a53462655e86ff1e7c64cb8a456"} Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.649053 4957 generic.go:334] "Generic (PLEG): container finished" podID="3050ec84-871a-460d-b50b-e52c87732f4d" containerID="9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f" exitCode=0 Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.649913 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-547bff997b-g5msk" event={"ID":"3050ec84-871a-460d-b50b-e52c87732f4d","Type":"ContainerDied","Data":"9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f"} Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.656205 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.840983 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.924921 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zls57"] Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.925216 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-zls57" podUID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerName="dnsmasq-dns" containerID="cri-o://2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986" gracePeriod=10 Dec 06 06:03:15 crc kubenswrapper[4957]: I1206 06:03:15.992813 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.272283 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.469759 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.485135 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.637950 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-config\") pod \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.638034 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-svc\") pod \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.638202 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-swift-storage-0\") pod \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.638231 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-sb\") pod \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.638336 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8xhn\" (UniqueName: \"kubernetes.io/projected/7859baf4-005a-4dad-8670-5d0ddc28c6e9-kube-api-access-h8xhn\") pod \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.638394 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-nb\") pod \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\" (UID: \"7859baf4-005a-4dad-8670-5d0ddc28c6e9\") " Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.652199 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7859baf4-005a-4dad-8670-5d0ddc28c6e9-kube-api-access-h8xhn" (OuterVolumeSpecName: "kube-api-access-h8xhn") pod "7859baf4-005a-4dad-8670-5d0ddc28c6e9" (UID: "7859baf4-005a-4dad-8670-5d0ddc28c6e9"). InnerVolumeSpecName "kube-api-access-h8xhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.740051 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8xhn\" (UniqueName: \"kubernetes.io/projected/7859baf4-005a-4dad-8670-5d0ddc28c6e9-kube-api-access-h8xhn\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.752584 4957 generic.go:334] "Generic (PLEG): container finished" podID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerID="2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986" exitCode=0 Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.752861 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zls57" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.754589 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.754574012 podStartE2EDuration="4.754574012s" podCreationTimestamp="2025-12-06 06:03:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:16.750431559 +0000 UTC m=+1436.400699181" watchObservedRunningTime="2025-12-06 06:03:16.754574012 +0000 UTC m=+1436.404841644" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.777894 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7859baf4-005a-4dad-8670-5d0ddc28c6e9" (UID: "7859baf4-005a-4dad-8670-5d0ddc28c6e9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.777921 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-config" (OuterVolumeSpecName: "config") pod "7859baf4-005a-4dad-8670-5d0ddc28c6e9" (UID: "7859baf4-005a-4dad-8670-5d0ddc28c6e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.782871 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7859baf4-005a-4dad-8670-5d0ddc28c6e9" (UID: "7859baf4-005a-4dad-8670-5d0ddc28c6e9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.782988 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c21557b2-c3ad-497a-ad42-ca81fff434bf","Type":"ContainerStarted","Data":"124919e2eb9be3497866e1e7591bd295738f04247814d724b2a48e6732d77fab"} Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.783037 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zls57" event={"ID":"7859baf4-005a-4dad-8670-5d0ddc28c6e9","Type":"ContainerDied","Data":"2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986"} Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.783066 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.783085 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.783098 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zls57" event={"ID":"7859baf4-005a-4dad-8670-5d0ddc28c6e9","Type":"ContainerDied","Data":"cea40f5f52a4243ac0e166617d275b12f35bbd8592313be1a8f09507c347a9f1"} Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.783112 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerStarted","Data":"2a130cc284505771a4431c9c5da0922c3e4329fff803d5e9b6a208408bca1586"} Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.783143 4957 scope.go:117] "RemoveContainer" containerID="2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.788386 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7859baf4-005a-4dad-8670-5d0ddc28c6e9" (UID: "7859baf4-005a-4dad-8670-5d0ddc28c6e9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.820179 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7859baf4-005a-4dad-8670-5d0ddc28c6e9" (UID: "7859baf4-005a-4dad-8670-5d0ddc28c6e9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.823644 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.831478 4957 scope.go:117] "RemoveContainer" containerID="93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.836672 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.637601153 podStartE2EDuration="6.836651861s" podCreationTimestamp="2025-12-06 06:03:10 +0000 UTC" firstStartedPulling="2025-12-06 06:03:11.427662931 +0000 UTC m=+1431.077930563" lastFinishedPulling="2025-12-06 06:03:15.626713639 +0000 UTC m=+1435.276981271" observedRunningTime="2025-12-06 06:03:16.832002725 +0000 UTC m=+1436.482270367" watchObservedRunningTime="2025-12-06 06:03:16.836651861 +0000 UTC m=+1436.486919493" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.842000 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.842039 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.842049 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.842059 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.842077 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7859baf4-005a-4dad-8670-5d0ddc28c6e9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.878049 4957 scope.go:117] "RemoveContainer" containerID="2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986" Dec 06 06:03:16 crc kubenswrapper[4957]: E1206 06:03:16.878634 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986\": container with ID starting with 2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986 not found: ID does not exist" containerID="2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.879041 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986"} err="failed to get container status \"2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986\": rpc error: code = NotFound desc = could not find container \"2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986\": container with ID starting with 2e125d55fd80d60ed0ba633a3bbef8563bcdb770a4ce77c455fd847c51ac6986 not found: ID does not exist" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.879110 4957 scope.go:117] "RemoveContainer" containerID="93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6" Dec 06 06:03:16 crc kubenswrapper[4957]: E1206 06:03:16.881069 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6\": container with ID starting with 93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6 not found: ID does not exist" containerID="93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6" Dec 06 06:03:16 crc kubenswrapper[4957]: I1206 06:03:16.881104 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6"} err="failed to get container status \"93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6\": rpc error: code = NotFound desc = could not find container \"93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6\": container with ID starting with 93cadb4da4526577dcd9242ee7c8e9d437a7360f721431035b9c9df2fa3846e6 not found: ID does not exist" Dec 06 06:03:17 crc kubenswrapper[4957]: I1206 06:03:17.103217 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zls57"] Dec 06 06:03:17 crc kubenswrapper[4957]: I1206 06:03:17.109769 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zls57"] Dec 06 06:03:17 crc kubenswrapper[4957]: I1206 06:03:17.789056 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="cinder-scheduler" containerID="cri-o://d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd" gracePeriod=30 Dec 06 06:03:17 crc kubenswrapper[4957]: I1206 06:03:17.789524 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="probe" containerID="cri-o://60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87" gracePeriod=30 Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.235855 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.379186 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-ovndb-tls-certs\") pod \"3050ec84-871a-460d-b50b-e52c87732f4d\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.379294 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-combined-ca-bundle\") pod \"3050ec84-871a-460d-b50b-e52c87732f4d\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.379357 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57cn8\" (UniqueName: \"kubernetes.io/projected/3050ec84-871a-460d-b50b-e52c87732f4d-kube-api-access-57cn8\") pod \"3050ec84-871a-460d-b50b-e52c87732f4d\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.379439 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-httpd-config\") pod \"3050ec84-871a-460d-b50b-e52c87732f4d\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.379485 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-config\") pod \"3050ec84-871a-460d-b50b-e52c87732f4d\" (UID: \"3050ec84-871a-460d-b50b-e52c87732f4d\") " Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.386994 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "3050ec84-871a-460d-b50b-e52c87732f4d" (UID: "3050ec84-871a-460d-b50b-e52c87732f4d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.390044 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3050ec84-871a-460d-b50b-e52c87732f4d-kube-api-access-57cn8" (OuterVolumeSpecName: "kube-api-access-57cn8") pod "3050ec84-871a-460d-b50b-e52c87732f4d" (UID: "3050ec84-871a-460d-b50b-e52c87732f4d"). InnerVolumeSpecName "kube-api-access-57cn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.440656 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-config" (OuterVolumeSpecName: "config") pod "3050ec84-871a-460d-b50b-e52c87732f4d" (UID: "3050ec84-871a-460d-b50b-e52c87732f4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.441090 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3050ec84-871a-460d-b50b-e52c87732f4d" (UID: "3050ec84-871a-460d-b50b-e52c87732f4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.462594 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "3050ec84-871a-460d-b50b-e52c87732f4d" (UID: "3050ec84-871a-460d-b50b-e52c87732f4d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.482271 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.482304 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57cn8\" (UniqueName: \"kubernetes.io/projected/3050ec84-871a-460d-b50b-e52c87732f4d-kube-api-access-57cn8\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.482321 4957 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.482333 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.482347 4957 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3050ec84-871a-460d-b50b-e52c87732f4d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.671012 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" path="/var/lib/kubelet/pods/7859baf4-005a-4dad-8670-5d0ddc28c6e9/volumes" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.789124 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-76c6cdbd84-nmkc9" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.801504 4957 generic.go:334] "Generic (PLEG): container finished" podID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerID="60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87" exitCode=0 Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.801588 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee976330-e8f0-4e33-9374-bbce8c1064f6","Type":"ContainerDied","Data":"60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87"} Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.803733 4957 generic.go:334] "Generic (PLEG): container finished" podID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerID="da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b" exitCode=0 Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.803802 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c6cdbd84-nmkc9" event={"ID":"3d4332b6-8252-435b-9a1c-bc3f27addaea","Type":"ContainerDied","Data":"da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b"} Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.806047 4957 generic.go:334] "Generic (PLEG): container finished" podID="3050ec84-871a-460d-b50b-e52c87732f4d" containerID="5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00" exitCode=0 Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.806077 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-547bff997b-g5msk" event={"ID":"3050ec84-871a-460d-b50b-e52c87732f4d","Type":"ContainerDied","Data":"5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00"} Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.806093 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-547bff997b-g5msk" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.806110 4957 scope.go:117] "RemoveContainer" containerID="9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.806097 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-547bff997b-g5msk" event={"ID":"3050ec84-871a-460d-b50b-e52c87732f4d","Type":"ContainerDied","Data":"c41136494c1847077362e62b4e6eff212bf68f3610022d04ae3529c9e66727f2"} Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.836820 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-547bff997b-g5msk"] Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.842770 4957 scope.go:117] "RemoveContainer" containerID="5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.846026 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-547bff997b-g5msk"] Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.869297 4957 scope.go:117] "RemoveContainer" containerID="9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f" Dec 06 06:03:18 crc kubenswrapper[4957]: E1206 06:03:18.869676 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f\": container with ID starting with 9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f not found: ID does not exist" containerID="9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.869705 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f"} err="failed to get container status \"9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f\": rpc error: code = NotFound desc = could not find container \"9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f\": container with ID starting with 9585400b77623003d099559dcc8dbcb9e8af3eafa40e2505d1247b79a2b43e4f not found: ID does not exist" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.869736 4957 scope.go:117] "RemoveContainer" containerID="5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00" Dec 06 06:03:18 crc kubenswrapper[4957]: E1206 06:03:18.870392 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00\": container with ID starting with 5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00 not found: ID does not exist" containerID="5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00" Dec 06 06:03:18 crc kubenswrapper[4957]: I1206 06:03:18.870425 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00"} err="failed to get container status \"5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00\": rpc error: code = NotFound desc = could not find container \"5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00\": container with ID starting with 5aa91ed6de799c8a3f4c5a5df2d27d261bd79915635b852bbb67de9bd4a3ab00 not found: ID does not exist" Dec 06 06:03:20 crc kubenswrapper[4957]: I1206 06:03:20.688421 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" path="/var/lib/kubelet/pods/3050ec84-871a-460d-b50b-e52c87732f4d/volumes" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.296877 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.458532 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data-custom\") pod \"ee976330-e8f0-4e33-9374-bbce8c1064f6\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.458575 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data\") pod \"ee976330-e8f0-4e33-9374-bbce8c1064f6\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.458630 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee976330-e8f0-4e33-9374-bbce8c1064f6-etc-machine-id\") pod \"ee976330-e8f0-4e33-9374-bbce8c1064f6\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.458761 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-combined-ca-bundle\") pod \"ee976330-e8f0-4e33-9374-bbce8c1064f6\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.458799 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-scripts\") pod \"ee976330-e8f0-4e33-9374-bbce8c1064f6\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.458868 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b7dv\" (UniqueName: \"kubernetes.io/projected/ee976330-e8f0-4e33-9374-bbce8c1064f6-kube-api-access-8b7dv\") pod \"ee976330-e8f0-4e33-9374-bbce8c1064f6\" (UID: \"ee976330-e8f0-4e33-9374-bbce8c1064f6\") " Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.459055 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee976330-e8f0-4e33-9374-bbce8c1064f6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ee976330-e8f0-4e33-9374-bbce8c1064f6" (UID: "ee976330-e8f0-4e33-9374-bbce8c1064f6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.459337 4957 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee976330-e8f0-4e33-9374-bbce8c1064f6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.467818 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-scripts" (OuterVolumeSpecName: "scripts") pod "ee976330-e8f0-4e33-9374-bbce8c1064f6" (UID: "ee976330-e8f0-4e33-9374-bbce8c1064f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.475105 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ee976330-e8f0-4e33-9374-bbce8c1064f6" (UID: "ee976330-e8f0-4e33-9374-bbce8c1064f6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.481278 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee976330-e8f0-4e33-9374-bbce8c1064f6-kube-api-access-8b7dv" (OuterVolumeSpecName: "kube-api-access-8b7dv") pod "ee976330-e8f0-4e33-9374-bbce8c1064f6" (UID: "ee976330-e8f0-4e33-9374-bbce8c1064f6"). InnerVolumeSpecName "kube-api-access-8b7dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.537062 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee976330-e8f0-4e33-9374-bbce8c1064f6" (UID: "ee976330-e8f0-4e33-9374-bbce8c1064f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.561887 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b7dv\" (UniqueName: \"kubernetes.io/projected/ee976330-e8f0-4e33-9374-bbce8c1064f6-kube-api-access-8b7dv\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.561924 4957 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.561933 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.561953 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.569339 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data" (OuterVolumeSpecName: "config-data") pod "ee976330-e8f0-4e33-9374-bbce8c1064f6" (UID: "ee976330-e8f0-4e33-9374-bbce8c1064f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.669486 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee976330-e8f0-4e33-9374-bbce8c1064f6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.855192 4957 generic.go:334] "Generic (PLEG): container finished" podID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerID="d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd" exitCode=0 Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.855245 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee976330-e8f0-4e33-9374-bbce8c1064f6","Type":"ContainerDied","Data":"d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd"} Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.855272 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.855293 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee976330-e8f0-4e33-9374-bbce8c1064f6","Type":"ContainerDied","Data":"81b63e435cb24271cd9bf2d21b2ab8cec588930d77430ba7baed1608c7a639a0"} Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.855314 4957 scope.go:117] "RemoveContainer" containerID="60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.879021 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.893400 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.898627 4957 scope.go:117] "RemoveContainer" containerID="d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.920510 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921103 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerName="dnsmasq-dns" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921128 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerName="dnsmasq-dns" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921140 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-api" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921176 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-api" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921199 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-httpd" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921206 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-httpd" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921222 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="cinder-scheduler" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921230 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="cinder-scheduler" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921243 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921250 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921265 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon-log" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921273 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon-log" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921289 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerName="init" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921297 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerName="init" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.921318 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="probe" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921326 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="probe" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921565 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921581 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="430ef178-b9c0-4759-979b-92620ceca29f" containerName="horizon-log" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921597 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="cinder-scheduler" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921611 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-httpd" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921629 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="7859baf4-005a-4dad-8670-5d0ddc28c6e9" containerName="dnsmasq-dns" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921645 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" containerName="probe" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.921657 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3050ec84-871a-460d-b50b-e52c87732f4d" containerName="neutron-api" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.922919 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.926038 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.945281 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.981006 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-config-data\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.981109 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.981144 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn5r6\" (UniqueName: \"kubernetes.io/projected/78ea9b70-3dde-4183-9ce4-c61326745bf2-kube-api-access-mn5r6\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.981412 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ea9b70-3dde-4183-9ce4-c61326745bf2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.981488 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-scripts\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.981630 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.987638 4957 scope.go:117] "RemoveContainer" containerID="60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.991470 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87\": container with ID starting with 60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87 not found: ID does not exist" containerID="60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.991509 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87"} err="failed to get container status \"60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87\": rpc error: code = NotFound desc = could not find container \"60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87\": container with ID starting with 60f9bb55932bb5b7e9e074f682d1c8c3b6a6dab4a5d11bd34277338536c93f87 not found: ID does not exist" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.991535 4957 scope.go:117] "RemoveContainer" containerID="d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd" Dec 06 06:03:22 crc kubenswrapper[4957]: E1206 06:03:22.992457 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd\": container with ID starting with d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd not found: ID does not exist" containerID="d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd" Dec 06 06:03:22 crc kubenswrapper[4957]: I1206 06:03:22.992572 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd"} err="failed to get container status \"d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd\": rpc error: code = NotFound desc = could not find container \"d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd\": container with ID starting with d1dcd2c61103d633dae833315c18cc3d54a8353f91a66aab00c3836fc23ae6cd not found: ID does not exist" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.084364 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.085327 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-config-data\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.085463 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.085497 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn5r6\" (UniqueName: \"kubernetes.io/projected/78ea9b70-3dde-4183-9ce4-c61326745bf2-kube-api-access-mn5r6\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.085666 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ea9b70-3dde-4183-9ce4-c61326745bf2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.085745 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-scripts\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.086223 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ea9b70-3dde-4183-9ce4-c61326745bf2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.090937 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.090967 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.091456 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-scripts\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.106365 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ea9b70-3dde-4183-9ce4-c61326745bf2-config-data\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.106825 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn5r6\" (UniqueName: \"kubernetes.io/projected/78ea9b70-3dde-4183-9ce4-c61326745bf2-kube-api-access-mn5r6\") pod \"cinder-scheduler-0\" (UID: \"78ea9b70-3dde-4183-9ce4-c61326745bf2\") " pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.288437 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.342984 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.567456 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c9f9fb958-p7bk5" Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.655289 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c9cb5bbd6-kzmqd"] Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.655973 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api-log" containerID="cri-o://fe09a4774559c1856570653c25fc450bcc3e78abc5d218f3d01920b10b355fe1" gracePeriod=30 Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.656454 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api" containerID="cri-o://029139da834c4cb80541a8869fd993103c7108e8258fb58a7be99771eb2a5dbc" gracePeriod=30 Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.844391 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.874657 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"78ea9b70-3dde-4183-9ce4-c61326745bf2","Type":"ContainerStarted","Data":"36d7b3b17eaf8169af4746ba809bff40e8771954d7e61a6751d382fe3563c41d"} Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.899695 4957 generic.go:334] "Generic (PLEG): container finished" podID="a9159107-1504-4952-bf38-703c55a5a029" containerID="fe09a4774559c1856570653c25fc450bcc3e78abc5d218f3d01920b10b355fe1" exitCode=143 Dec 06 06:03:23 crc kubenswrapper[4957]: I1206 06:03:23.899788 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" event={"ID":"a9159107-1504-4952-bf38-703c55a5a029","Type":"ContainerDied","Data":"fe09a4774559c1856570653c25fc450bcc3e78abc5d218f3d01920b10b355fe1"} Dec 06 06:03:24 crc kubenswrapper[4957]: I1206 06:03:24.677673 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee976330-e8f0-4e33-9374-bbce8c1064f6" path="/var/lib/kubelet/pods/ee976330-e8f0-4e33-9374-bbce8c1064f6/volumes" Dec 06 06:03:24 crc kubenswrapper[4957]: I1206 06:03:24.949267 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"78ea9b70-3dde-4183-9ce4-c61326745bf2","Type":"ContainerStarted","Data":"04ca37680b637faf285868ef2e4b1c12fa83be95d1639c900cf029b8105cb15b"} Dec 06 06:03:25 crc kubenswrapper[4957]: I1206 06:03:25.958059 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"78ea9b70-3dde-4183-9ce4-c61326745bf2","Type":"ContainerStarted","Data":"cbc5a1937ff3dd0a1c20e697da72ca5fc2509da584ffa93d6c60b5d1585d1154"} Dec 06 06:03:25 crc kubenswrapper[4957]: I1206 06:03:25.985502 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.985479982 podStartE2EDuration="3.985479982s" podCreationTimestamp="2025-12-06 06:03:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:25.979293444 +0000 UTC m=+1445.629561096" watchObservedRunningTime="2025-12-06 06:03:25.985479982 +0000 UTC m=+1445.635747614" Dec 06 06:03:26 crc kubenswrapper[4957]: I1206 06:03:26.023334 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 06:03:26 crc kubenswrapper[4957]: I1206 06:03:26.440106 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7c7fdbc8b-njnq6" Dec 06 06:03:26 crc kubenswrapper[4957]: I1206 06:03:26.858969 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:50320->10.217.0.159:9311: read: connection reset by peer" Dec 06 06:03:26 crc kubenswrapper[4957]: I1206 06:03:26.859001 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:50304->10.217.0.159:9311: read: connection reset by peer" Dec 06 06:03:26 crc kubenswrapper[4957]: I1206 06:03:26.998175 4957 generic.go:334] "Generic (PLEG): container finished" podID="a9159107-1504-4952-bf38-703c55a5a029" containerID="029139da834c4cb80541a8869fd993103c7108e8258fb58a7be99771eb2a5dbc" exitCode=0 Dec 06 06:03:26 crc kubenswrapper[4957]: I1206 06:03:26.999647 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" event={"ID":"a9159107-1504-4952-bf38-703c55a5a029","Type":"ContainerDied","Data":"029139da834c4cb80541a8869fd993103c7108e8258fb58a7be99771eb2a5dbc"} Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.263766 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.406157 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvm9w\" (UniqueName: \"kubernetes.io/projected/a9159107-1504-4952-bf38-703c55a5a029-kube-api-access-tvm9w\") pod \"a9159107-1504-4952-bf38-703c55a5a029\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.406326 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-combined-ca-bundle\") pod \"a9159107-1504-4952-bf38-703c55a5a029\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.406448 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9159107-1504-4952-bf38-703c55a5a029-logs\") pod \"a9159107-1504-4952-bf38-703c55a5a029\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.406519 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data\") pod \"a9159107-1504-4952-bf38-703c55a5a029\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.406589 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data-custom\") pod \"a9159107-1504-4952-bf38-703c55a5a029\" (UID: \"a9159107-1504-4952-bf38-703c55a5a029\") " Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.407247 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9159107-1504-4952-bf38-703c55a5a029-logs" (OuterVolumeSpecName: "logs") pod "a9159107-1504-4952-bf38-703c55a5a029" (UID: "a9159107-1504-4952-bf38-703c55a5a029"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.412313 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a9159107-1504-4952-bf38-703c55a5a029" (UID: "a9159107-1504-4952-bf38-703c55a5a029"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.418986 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9159107-1504-4952-bf38-703c55a5a029-kube-api-access-tvm9w" (OuterVolumeSpecName: "kube-api-access-tvm9w") pod "a9159107-1504-4952-bf38-703c55a5a029" (UID: "a9159107-1504-4952-bf38-703c55a5a029"). InnerVolumeSpecName "kube-api-access-tvm9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.442949 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9159107-1504-4952-bf38-703c55a5a029" (UID: "a9159107-1504-4952-bf38-703c55a5a029"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.462812 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data" (OuterVolumeSpecName: "config-data") pod "a9159107-1504-4952-bf38-703c55a5a029" (UID: "a9159107-1504-4952-bf38-703c55a5a029"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.508100 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvm9w\" (UniqueName: \"kubernetes.io/projected/a9159107-1504-4952-bf38-703c55a5a029-kube-api-access-tvm9w\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.508132 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.508143 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9159107-1504-4952-bf38-703c55a5a029-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.508152 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.508163 4957 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9159107-1504-4952-bf38-703c55a5a029-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.511440 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 06:03:27 crc kubenswrapper[4957]: E1206 06:03:27.511805 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api-log" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.511823 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api-log" Dec 06 06:03:27 crc kubenswrapper[4957]: E1206 06:03:27.511853 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.511861 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.512052 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api-log" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.512066 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9159107-1504-4952-bf38-703c55a5a029" containerName="barbican-api" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.512811 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.515616 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.515797 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.515943 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jn9b6" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.518300 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.710461 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-openstack-config-secret\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.710503 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmgk\" (UniqueName: \"kubernetes.io/projected/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-kube-api-access-8cmgk\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.711307 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-openstack-config\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.711372 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.813344 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-openstack-config\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.813777 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.813872 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-openstack-config-secret\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.813927 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmgk\" (UniqueName: \"kubernetes.io/projected/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-kube-api-access-8cmgk\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.814397 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-openstack-config\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.818140 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.818212 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-openstack-config-secret\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.834350 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmgk\" (UniqueName: \"kubernetes.io/projected/663bc0fb-e06b-4c91-8a36-a97d32fa63f3-kube-api-access-8cmgk\") pod \"openstackclient\" (UID: \"663bc0fb-e06b-4c91-8a36-a97d32fa63f3\") " pod="openstack/openstackclient" Dec 06 06:03:27 crc kubenswrapper[4957]: I1206 06:03:27.863437 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.022700 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" event={"ID":"a9159107-1504-4952-bf38-703c55a5a029","Type":"ContainerDied","Data":"6891e471f722e982c27ca8ada1660d09aea020a360b1470b915baf73e261fda0"} Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.022759 4957 scope.go:117] "RemoveContainer" containerID="029139da834c4cb80541a8869fd993103c7108e8258fb58a7be99771eb2a5dbc" Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.022985 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9cb5bbd6-kzmqd" Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.056992 4957 scope.go:117] "RemoveContainer" containerID="fe09a4774559c1856570653c25fc450bcc3e78abc5d218f3d01920b10b355fe1" Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.078891 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-c9cb5bbd6-kzmqd"] Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.087941 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-c9cb5bbd6-kzmqd"] Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.288766 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.456796 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.672538 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9159107-1504-4952-bf38-703c55a5a029" path="/var/lib/kubelet/pods/a9159107-1504-4952-bf38-703c55a5a029/volumes" Dec 06 06:03:28 crc kubenswrapper[4957]: I1206 06:03:28.789041 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-76c6cdbd84-nmkc9" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 06:03:29 crc kubenswrapper[4957]: I1206 06:03:29.040129 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"663bc0fb-e06b-4c91-8a36-a97d32fa63f3","Type":"ContainerStarted","Data":"dac7f5202ffa728ed497f05b35b5347dbf9accaaccce04422ad123aace6089ca"} Dec 06 06:03:29 crc kubenswrapper[4957]: I1206 06:03:29.899323 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:03:29 crc kubenswrapper[4957]: I1206 06:03:29.903868 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69dd5567fb-lm2jl" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.240858 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5789568449-4t258"] Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.242870 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.250282 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.250395 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.250498 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.260574 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5789568449-4t258"] Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.320533 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/475d23a9-0dda-4ba5-8af9-108574997a4f-etc-swift\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.320578 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475d23a9-0dda-4ba5-8af9-108574997a4f-run-httpd\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.320612 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-public-tls-certs\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.320805 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475d23a9-0dda-4ba5-8af9-108574997a4f-log-httpd\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.320895 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-internal-tls-certs\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.320940 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-combined-ca-bundle\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.321020 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-config-data\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.321242 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67xrt\" (UniqueName: \"kubernetes.io/projected/475d23a9-0dda-4ba5-8af9-108574997a4f-kube-api-access-67xrt\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423239 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67xrt\" (UniqueName: \"kubernetes.io/projected/475d23a9-0dda-4ba5-8af9-108574997a4f-kube-api-access-67xrt\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423317 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/475d23a9-0dda-4ba5-8af9-108574997a4f-etc-swift\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423343 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475d23a9-0dda-4ba5-8af9-108574997a4f-run-httpd\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423369 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-public-tls-certs\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423432 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475d23a9-0dda-4ba5-8af9-108574997a4f-log-httpd\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423456 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-internal-tls-certs\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423481 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-combined-ca-bundle\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.423523 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-config-data\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.424034 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475d23a9-0dda-4ba5-8af9-108574997a4f-log-httpd\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.424720 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/475d23a9-0dda-4ba5-8af9-108574997a4f-run-httpd\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.430076 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-internal-tls-certs\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.431717 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-config-data\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.434935 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/475d23a9-0dda-4ba5-8af9-108574997a4f-etc-swift\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.439292 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-combined-ca-bundle\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.439376 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/475d23a9-0dda-4ba5-8af9-108574997a4f-public-tls-certs\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.448620 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67xrt\" (UniqueName: \"kubernetes.io/projected/475d23a9-0dda-4ba5-8af9-108574997a4f-kube-api-access-67xrt\") pod \"swift-proxy-5789568449-4t258\" (UID: \"475d23a9-0dda-4ba5-8af9-108574997a4f\") " pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:32 crc kubenswrapper[4957]: I1206 06:03:32.562375 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:33 crc kubenswrapper[4957]: I1206 06:03:33.163506 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5789568449-4t258"] Dec 06 06:03:33 crc kubenswrapper[4957]: I1206 06:03:33.514929 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:03:33 crc kubenswrapper[4957]: I1206 06:03:33.515491 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-log" containerID="cri-o://85679ee7f099f70cfabd5b1d2aeb68997468664d5994775abff132ffa150a703" gracePeriod=30 Dec 06 06:03:33 crc kubenswrapper[4957]: I1206 06:03:33.515942 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-httpd" containerID="cri-o://0a057a094f04b6b804a78b44a8c9aa245162822d6f56d0aa83222484239b62ea" gracePeriod=30 Dec 06 06:03:33 crc kubenswrapper[4957]: I1206 06:03:33.651206 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.104272 4957 generic.go:334] "Generic (PLEG): container finished" podID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerID="85679ee7f099f70cfabd5b1d2aeb68997468664d5994775abff132ffa150a703" exitCode=143 Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.104511 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"376b5f5c-4e23-4e75-9619-698a4d95eab5","Type":"ContainerDied","Data":"85679ee7f099f70cfabd5b1d2aeb68997468664d5994775abff132ffa150a703"} Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.106121 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5789568449-4t258" event={"ID":"475d23a9-0dda-4ba5-8af9-108574997a4f","Type":"ContainerStarted","Data":"48eab4d8c208f90d77c3e394b12426539f68d5b8a19b90e562a1787132b3fefb"} Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.106149 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5789568449-4t258" event={"ID":"475d23a9-0dda-4ba5-8af9-108574997a4f","Type":"ContainerStarted","Data":"98faeac2b8250ed294c70fcbc664baedc70188be9715de05dca380af2ede6927"} Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.106161 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5789568449-4t258" event={"ID":"475d23a9-0dda-4ba5-8af9-108574997a4f","Type":"ContainerStarted","Data":"2e5d246e42107d39687cfd9a285d20c6b3eb4e4749eee4f8f77f140041d2e7d0"} Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.106378 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.106671 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.136597 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5789568449-4t258" podStartSLOduration=2.136575249 podStartE2EDuration="2.136575249s" podCreationTimestamp="2025-12-06 06:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:34.13329725 +0000 UTC m=+1453.783564882" watchObservedRunningTime="2025-12-06 06:03:34.136575249 +0000 UTC m=+1453.786842881" Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.174683 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.178411 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-central-agent" containerID="cri-o://f9ea6af81eeac4d2567ee02d47f19d6b17a9911b3df146d3d748f1c89739bcf7" gracePeriod=30 Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.180064 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="sg-core" containerID="cri-o://c67ec980d4e6c56f4805e1d461b1e25fdf1bbe6c8597c01c9000e4d5a83bf411" gracePeriod=30 Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.180196 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="proxy-httpd" containerID="cri-o://2a130cc284505771a4431c9c5da0922c3e4329fff803d5e9b6a208408bca1586" gracePeriod=30 Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.180214 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-notification-agent" containerID="cri-o://75a254a9e897fcbd2463c3b029ba7056131f1bd241f50428f75ca10c32031549" gracePeriod=30 Dec 06 06:03:34 crc kubenswrapper[4957]: I1206 06:03:34.298264 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.163:3000/\": EOF" Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.126917 4957 generic.go:334] "Generic (PLEG): container finished" podID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerID="2a130cc284505771a4431c9c5da0922c3e4329fff803d5e9b6a208408bca1586" exitCode=0 Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.126953 4957 generic.go:334] "Generic (PLEG): container finished" podID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerID="c67ec980d4e6c56f4805e1d461b1e25fdf1bbe6c8597c01c9000e4d5a83bf411" exitCode=2 Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.126962 4957 generic.go:334] "Generic (PLEG): container finished" podID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerID="f9ea6af81eeac4d2567ee02d47f19d6b17a9911b3df146d3d748f1c89739bcf7" exitCode=0 Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.127903 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerDied","Data":"2a130cc284505771a4431c9c5da0922c3e4329fff803d5e9b6a208408bca1586"} Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.127932 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerDied","Data":"c67ec980d4e6c56f4805e1d461b1e25fdf1bbe6c8597c01c9000e4d5a83bf411"} Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.127942 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerDied","Data":"f9ea6af81eeac4d2567ee02d47f19d6b17a9911b3df146d3d748f1c89739bcf7"} Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.657032 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.657712 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-httpd" containerID="cri-o://54444629802b56358878379ddd680c333de89593fda09553e8fe24d76d03288b" gracePeriod=30 Dec 06 06:03:35 crc kubenswrapper[4957]: I1206 06:03:35.657580 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-log" containerID="cri-o://4fe6ed5c1695b975e31be44f6177795022263bc46dadc34ae744b6e40adb918d" gracePeriod=30 Dec 06 06:03:36 crc kubenswrapper[4957]: I1206 06:03:36.141186 4957 generic.go:334] "Generic (PLEG): container finished" podID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerID="4fe6ed5c1695b975e31be44f6177795022263bc46dadc34ae744b6e40adb918d" exitCode=143 Dec 06 06:03:36 crc kubenswrapper[4957]: I1206 06:03:36.141238 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1085fe7-6d61-4e23-85ba-51ec7740b8c4","Type":"ContainerDied","Data":"4fe6ed5c1695b975e31be44f6177795022263bc46dadc34ae744b6e40adb918d"} Dec 06 06:03:36 crc kubenswrapper[4957]: I1206 06:03:36.755247 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.148:9292/healthcheck\": dial tcp 10.217.0.148:9292: connect: connection refused" Dec 06 06:03:36 crc kubenswrapper[4957]: I1206 06:03:36.755627 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.148:9292/healthcheck\": dial tcp 10.217.0.148:9292: connect: connection refused" Dec 06 06:03:37 crc kubenswrapper[4957]: I1206 06:03:37.161228 4957 generic.go:334] "Generic (PLEG): container finished" podID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerID="0a057a094f04b6b804a78b44a8c9aa245162822d6f56d0aa83222484239b62ea" exitCode=0 Dec 06 06:03:37 crc kubenswrapper[4957]: I1206 06:03:37.161301 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"376b5f5c-4e23-4e75-9619-698a4d95eab5","Type":"ContainerDied","Data":"0a057a094f04b6b804a78b44a8c9aa245162822d6f56d0aa83222484239b62ea"} Dec 06 06:03:38 crc kubenswrapper[4957]: I1206 06:03:38.789399 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-76c6cdbd84-nmkc9" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 06:03:38 crc kubenswrapper[4957]: I1206 06:03:38.789819 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:03:39 crc kubenswrapper[4957]: I1206 06:03:39.193993 4957 generic.go:334] "Generic (PLEG): container finished" podID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerID="54444629802b56358878379ddd680c333de89593fda09553e8fe24d76d03288b" exitCode=0 Dec 06 06:03:39 crc kubenswrapper[4957]: I1206 06:03:39.194063 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1085fe7-6d61-4e23-85ba-51ec7740b8c4","Type":"ContainerDied","Data":"54444629802b56358878379ddd680c333de89593fda09553e8fe24d76d03288b"} Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.228562 4957 generic.go:334] "Generic (PLEG): container finished" podID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerID="75a254a9e897fcbd2463c3b029ba7056131f1bd241f50428f75ca10c32031549" exitCode=0 Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.228703 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerDied","Data":"75a254a9e897fcbd2463c3b029ba7056131f1bd241f50428f75ca10c32031549"} Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.322209 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.384696 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.393502 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402126 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-combined-ca-bundle\") pod \"b38c57a0-87fd-41ad-9108-98cc86622c56\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402177 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-scripts\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402236 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-log-httpd\") pod \"b38c57a0-87fd-41ad-9108-98cc86622c56\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402282 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-combined-ca-bundle\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402316 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-config-data\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402352 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn67g\" (UniqueName: \"kubernetes.io/projected/b38c57a0-87fd-41ad-9108-98cc86622c56-kube-api-access-fn67g\") pod \"b38c57a0-87fd-41ad-9108-98cc86622c56\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402402 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-public-tls-certs\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402429 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-scripts\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402453 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-run-httpd\") pod \"b38c57a0-87fd-41ad-9108-98cc86622c56\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402488 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402516 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-httpd-run\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402544 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-config-data\") pod \"b38c57a0-87fd-41ad-9108-98cc86622c56\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402568 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-combined-ca-bundle\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402593 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-scripts\") pod \"b38c57a0-87fd-41ad-9108-98cc86622c56\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402619 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-config-data\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402658 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-httpd-run\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402683 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g46t9\" (UniqueName: \"kubernetes.io/projected/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-kube-api-access-g46t9\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402710 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-sg-core-conf-yaml\") pod \"b38c57a0-87fd-41ad-9108-98cc86622c56\" (UID: \"b38c57a0-87fd-41ad-9108-98cc86622c56\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402753 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-logs\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402774 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402793 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-internal-tls-certs\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402822 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-logs\") pod \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\" (UID: \"a1085fe7-6d61-4e23-85ba-51ec7740b8c4\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.402859 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4thw\" (UniqueName: \"kubernetes.io/projected/376b5f5c-4e23-4e75-9619-698a4d95eab5-kube-api-access-n4thw\") pod \"376b5f5c-4e23-4e75-9619-698a4d95eab5\" (UID: \"376b5f5c-4e23-4e75-9619-698a4d95eab5\") " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.411503 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b38c57a0-87fd-41ad-9108-98cc86622c56" (UID: "b38c57a0-87fd-41ad-9108-98cc86622c56"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.413913 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b38c57a0-87fd-41ad-9108-98cc86622c56" (UID: "b38c57a0-87fd-41ad-9108-98cc86622c56"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.414132 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-logs" (OuterVolumeSpecName: "logs") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.414595 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-logs" (OuterVolumeSpecName: "logs") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.417631 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.421579 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-scripts" (OuterVolumeSpecName: "scripts") pod "b38c57a0-87fd-41ad-9108-98cc86622c56" (UID: "b38c57a0-87fd-41ad-9108-98cc86622c56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.422389 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.422885 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.424031 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.426439 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-kube-api-access-g46t9" (OuterVolumeSpecName: "kube-api-access-g46t9") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "kube-api-access-g46t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.426517 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b38c57a0-87fd-41ad-9108-98cc86622c56-kube-api-access-fn67g" (OuterVolumeSpecName: "kube-api-access-fn67g") pod "b38c57a0-87fd-41ad-9108-98cc86622c56" (UID: "b38c57a0-87fd-41ad-9108-98cc86622c56"). InnerVolumeSpecName "kube-api-access-fn67g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.426555 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376b5f5c-4e23-4e75-9619-698a4d95eab5-kube-api-access-n4thw" (OuterVolumeSpecName: "kube-api-access-n4thw") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "kube-api-access-n4thw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.438047 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-scripts" (OuterVolumeSpecName: "scripts") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.451168 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-scripts" (OuterVolumeSpecName: "scripts") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.498730 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b38c57a0-87fd-41ad-9108-98cc86622c56" (UID: "b38c57a0-87fd-41ad-9108-98cc86622c56"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506486 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506545 4957 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506560 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4thw\" (UniqueName: \"kubernetes.io/projected/376b5f5c-4e23-4e75-9619-698a4d95eab5-kube-api-access-n4thw\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506576 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506588 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506599 4957 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506610 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn67g\" (UniqueName: \"kubernetes.io/projected/b38c57a0-87fd-41ad-9108-98cc86622c56-kube-api-access-fn67g\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506624 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506635 4957 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b38c57a0-87fd-41ad-9108-98cc86622c56-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506957 4957 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506973 4957 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.506983 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.507011 4957 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/376b5f5c-4e23-4e75-9619-698a4d95eab5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.507022 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g46t9\" (UniqueName: \"kubernetes.io/projected/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-kube-api-access-g46t9\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.507032 4957 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.512296 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.514463 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-config-data" (OuterVolumeSpecName: "config-data") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.516344 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b38c57a0-87fd-41ad-9108-98cc86622c56" (UID: "b38c57a0-87fd-41ad-9108-98cc86622c56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.540198 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.542980 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1085fe7-6d61-4e23-85ba-51ec7740b8c4" (UID: "a1085fe7-6d61-4e23-85ba-51ec7740b8c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.548403 4957 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.548523 4957 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.552050 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-config-data" (OuterVolumeSpecName: "config-data") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.563995 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "376b5f5c-4e23-4e75-9619-698a4d95eab5" (UID: "376b5f5c-4e23-4e75-9619-698a4d95eab5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.582581 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-config-data" (OuterVolumeSpecName: "config-data") pod "b38c57a0-87fd-41ad-9108-98cc86622c56" (UID: "b38c57a0-87fd-41ad-9108-98cc86622c56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608474 4957 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608509 4957 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608521 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608529 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608538 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608547 4957 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/376b5f5c-4e23-4e75-9619-698a4d95eab5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608555 4957 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608562 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38c57a0-87fd-41ad-9108-98cc86622c56-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608569 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:40 crc kubenswrapper[4957]: I1206 06:03:40.608577 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1085fe7-6d61-4e23-85ba-51ec7740b8c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.241498 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.241429 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1085fe7-6d61-4e23-85ba-51ec7740b8c4","Type":"ContainerDied","Data":"87c72743a2df1a51d9015cfede0aee1c76428c106ec7d5fe729d4c01c8f06d2c"} Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.242075 4957 scope.go:117] "RemoveContainer" containerID="54444629802b56358878379ddd680c333de89593fda09553e8fe24d76d03288b" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.243230 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"663bc0fb-e06b-4c91-8a36-a97d32fa63f3","Type":"ContainerStarted","Data":"9385ce92522b2a3532a5dd62cf8eb3f459cef4b5a45c93fe943e0d11ca1eb52d"} Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.251046 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b38c57a0-87fd-41ad-9108-98cc86622c56","Type":"ContainerDied","Data":"49b54e089892aeb4c9b75c601621bc3beb3d4de162e49b28fb367fc0edec8252"} Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.251173 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.254763 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"376b5f5c-4e23-4e75-9619-698a4d95eab5","Type":"ContainerDied","Data":"dd745fa12b620cbfc218082405935e489c68f692fa141055e3118c2aba3a007a"} Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.255482 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.323671 4957 scope.go:117] "RemoveContainer" containerID="4fe6ed5c1695b975e31be44f6177795022263bc46dadc34ae744b6e40adb918d" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.326041 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.356961 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.395248 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.412342 4957 scope.go:117] "RemoveContainer" containerID="2a130cc284505771a4431c9c5da0922c3e4329fff803d5e9b6a208408bca1586" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.414169 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428108 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428531 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428556 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428573 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-notification-agent" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428582 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-notification-agent" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428592 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-central-agent" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428600 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-central-agent" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428629 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="proxy-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428637 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="proxy-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428661 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="sg-core" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428669 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="sg-core" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428683 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-log" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428691 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-log" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428708 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428715 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.428727 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-log" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428735 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-log" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.428994 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.429009 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-central-agent" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.429023 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-log" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.429034 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="proxy-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.429047 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" containerName="glance-log" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.429061 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="sg-core" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.429077 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" containerName="glance-httpd" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.429091 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" containerName="ceilometer-notification-agent" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.430220 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.432105 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.436156 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.436286 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.436350 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-prhgl" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.444375 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.460027 4957 scope.go:117] "RemoveContainer" containerID="c67ec980d4e6c56f4805e1d461b1e25fdf1bbe6c8597c01c9000e4d5a83bf411" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.463676 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.473327 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.481954 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.488210 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.490308 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.490502 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:03:41 crc kubenswrapper[4957]: E1206 06:03:41.493803 4957 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: , extraDiskErr: could not stat "/var/log/pods/openstack_glance-default-external-api-0_376b5f5c-4e23-4e75-9619-698a4d95eab5/glance-log/0.log" to get inode usage: stat /var/log/pods/openstack_glance-default-external-api-0_376b5f5c-4e23-4e75-9619-698a4d95eab5/glance-log/0.log: no such file or directory Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.496248 4957 scope.go:117] "RemoveContainer" containerID="75a254a9e897fcbd2463c3b029ba7056131f1bd241f50428f75ca10c32031549" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.499266 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.511459 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.521927 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.523237 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.523693 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.523987 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527566 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8deee48-5aef-4889-b7ab-b7f23144cbc8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527630 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527652 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527679 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527714 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527736 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8deee48-5aef-4889-b7ab-b7f23144cbc8-logs\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527788 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.527845 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzj4j\" (UniqueName: \"kubernetes.io/projected/e8deee48-5aef-4889-b7ab-b7f23144cbc8-kube-api-access-qzj4j\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.532282 4957 scope.go:117] "RemoveContainer" containerID="f9ea6af81eeac4d2567ee02d47f19d6b17a9911b3df146d3d748f1c89739bcf7" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.558350 4957 scope.go:117] "RemoveContainer" containerID="0a057a094f04b6b804a78b44a8c9aa245162822d6f56d0aa83222484239b62ea" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.581974 4957 scope.go:117] "RemoveContainer" containerID="85679ee7f099f70cfabd5b1d2aeb68997468664d5994775abff132ffa150a703" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.629672 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8deee48-5aef-4889-b7ab-b7f23144cbc8-logs\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.629741 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-config-data\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.629795 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.629815 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.629882 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-log-httpd\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.629910 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzj4j\" (UniqueName: \"kubernetes.io/projected/e8deee48-5aef-4889-b7ab-b7f23144cbc8-kube-api-access-qzj4j\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630110 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-scripts\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630177 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630246 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630325 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8deee48-5aef-4889-b7ab-b7f23144cbc8-logs\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630410 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2319a013-4227-4643-9b7f-d60b4d501af7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630419 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630437 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2319a013-4227-4643-9b7f-d60b4d501af7-logs\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630532 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8deee48-5aef-4889-b7ab-b7f23144cbc8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630572 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-run-httpd\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630623 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmtz9\" (UniqueName: \"kubernetes.io/projected/2319a013-4227-4643-9b7f-d60b4d501af7-kube-api-access-nmtz9\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630648 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-scripts\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630696 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhbwn\" (UniqueName: \"kubernetes.io/projected/f9f19718-7793-4223-83af-373ffe31a15b-kube-api-access-lhbwn\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630727 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-config-data\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630769 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630798 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630825 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630939 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630988 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.631008 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.630800 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e8deee48-5aef-4889-b7ab-b7f23144cbc8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.636942 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.638779 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.639865 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.650710 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8deee48-5aef-4889-b7ab-b7f23144cbc8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.654565 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzj4j\" (UniqueName: \"kubernetes.io/projected/e8deee48-5aef-4889-b7ab-b7f23144cbc8-kube-api-access-qzj4j\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.687111 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e8deee48-5aef-4889-b7ab-b7f23144cbc8\") " pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733004 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-config-data\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733103 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733166 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-log-httpd\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733247 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-scripts\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733265 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733302 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733332 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2319a013-4227-4643-9b7f-d60b4d501af7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733350 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2319a013-4227-4643-9b7f-d60b4d501af7-logs\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733393 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-run-httpd\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733410 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmtz9\" (UniqueName: \"kubernetes.io/projected/2319a013-4227-4643-9b7f-d60b4d501af7-kube-api-access-nmtz9\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733426 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-scripts\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733463 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhbwn\" (UniqueName: \"kubernetes.io/projected/f9f19718-7793-4223-83af-373ffe31a15b-kube-api-access-lhbwn\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733482 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-config-data\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733498 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733583 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.733710 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-log-httpd\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.734377 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-run-httpd\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.734744 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2319a013-4227-4643-9b7f-d60b4d501af7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.734846 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2319a013-4227-4643-9b7f-d60b4d501af7-logs\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.735173 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.740503 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-scripts\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.740653 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.741222 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.741466 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-config-data\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.741716 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2319a013-4227-4643-9b7f-d60b4d501af7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.741809 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.742005 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-config-data\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.745309 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-scripts\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.753657 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhbwn\" (UniqueName: \"kubernetes.io/projected/f9f19718-7793-4223-83af-373ffe31a15b-kube-api-access-lhbwn\") pod \"ceilometer-0\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.756291 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.756627 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmtz9\" (UniqueName: \"kubernetes.io/projected/2319a013-4227-4643-9b7f-d60b4d501af7-kube-api-access-nmtz9\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.768266 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"2319a013-4227-4643-9b7f-d60b4d501af7\") " pod="openstack/glance-default-external-api-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.807586 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:41 crc kubenswrapper[4957]: I1206 06:03:41.848079 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.310485 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.318740 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.814792444 podStartE2EDuration="15.318719675s" podCreationTimestamp="2025-12-06 06:03:27 +0000 UTC" firstStartedPulling="2025-12-06 06:03:28.471928915 +0000 UTC m=+1448.122196557" lastFinishedPulling="2025-12-06 06:03:39.975856156 +0000 UTC m=+1459.626123788" observedRunningTime="2025-12-06 06:03:42.306526036 +0000 UTC m=+1461.956793668" watchObservedRunningTime="2025-12-06 06:03:42.318719675 +0000 UTC m=+1461.968987307" Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.370157 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.455083 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.577549 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.579941 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5789568449-4t258" Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.683715 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="376b5f5c-4e23-4e75-9619-698a4d95eab5" path="/var/lib/kubelet/pods/376b5f5c-4e23-4e75-9619-698a4d95eab5/volumes" Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.689497 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1085fe7-6d61-4e23-85ba-51ec7740b8c4" path="/var/lib/kubelet/pods/a1085fe7-6d61-4e23-85ba-51ec7740b8c4/volumes" Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.690968 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b38c57a0-87fd-41ad-9108-98cc86622c56" path="/var/lib/kubelet/pods/b38c57a0-87fd-41ad-9108-98cc86622c56/volumes" Dec 06 06:03:42 crc kubenswrapper[4957]: I1206 06:03:42.746000 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 06:03:42 crc kubenswrapper[4957]: E1206 06:03:42.768931 4957 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/4807a6fcf837b20bf4df133161b7904022a4be132e5a18c4c9ae4442a5fdb234/diff" to get inode usage: stat /var/lib/containers/storage/overlay/4807a6fcf837b20bf4df133161b7904022a4be132e5a18c4c9ae4442a5fdb234/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_glance-default-external-api-0_376b5f5c-4e23-4e75-9619-698a4d95eab5/glance-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_glance-default-external-api-0_376b5f5c-4e23-4e75-9619-698a4d95eab5/glance-httpd/0.log: no such file or directory Dec 06 06:03:43 crc kubenswrapper[4957]: I1206 06:03:43.302827 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2319a013-4227-4643-9b7f-d60b4d501af7","Type":"ContainerStarted","Data":"8ddc9e62c80b21fe5d622645a5198c729c06b5ca7f7d77dde484fe761e7754e2"} Dec 06 06:03:43 crc kubenswrapper[4957]: I1206 06:03:43.306549 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerStarted","Data":"43897b185c4110875697ff4e028f414a136109785e14b5c97ff5f43993935600"} Dec 06 06:03:43 crc kubenswrapper[4957]: I1206 06:03:43.311488 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e8deee48-5aef-4889-b7ab-b7f23144cbc8","Type":"ContainerStarted","Data":"f7bbe71e62029b30d0492c73f5ba8f49097d96b9f6de1c2f619cfbce30d92e05"} Dec 06 06:03:43 crc kubenswrapper[4957]: I1206 06:03:43.311540 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e8deee48-5aef-4889-b7ab-b7f23144cbc8","Type":"ContainerStarted","Data":"c0cf9e45108b50285e3e0c5754824cb217d070f32e377f2ab4b5de8f508ebccc"} Dec 06 06:03:44 crc kubenswrapper[4957]: I1206 06:03:44.345277 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e8deee48-5aef-4889-b7ab-b7f23144cbc8","Type":"ContainerStarted","Data":"79a27db18bd04862e15bb7c331aca661c96bc58aa79e857d379fbb43138b2106"} Dec 06 06:03:44 crc kubenswrapper[4957]: I1206 06:03:44.355366 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2319a013-4227-4643-9b7f-d60b4d501af7","Type":"ContainerStarted","Data":"5691cd9cd099ad212edb64a631eaa4e020dfd3ae43bc054ff2d8efa0ddfcdcec"} Dec 06 06:03:44 crc kubenswrapper[4957]: I1206 06:03:44.357305 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerStarted","Data":"1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4"} Dec 06 06:03:44 crc kubenswrapper[4957]: I1206 06:03:44.373637 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.373619098 podStartE2EDuration="3.373619098s" podCreationTimestamp="2025-12-06 06:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:44.368042498 +0000 UTC m=+1464.018310150" watchObservedRunningTime="2025-12-06 06:03:44.373619098 +0000 UTC m=+1464.023886730" Dec 06 06:03:44 crc kubenswrapper[4957]: E1206 06:03:44.965647 4957 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d4332b6_8252_435b_9a1c_bc3f27addaea.slice/crio-conmon-1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.146174 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.214283 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-combined-ca-bundle\") pod \"3d4332b6-8252-435b-9a1c-bc3f27addaea\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.214817 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-secret-key\") pod \"3d4332b6-8252-435b-9a1c-bc3f27addaea\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.214939 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d4332b6-8252-435b-9a1c-bc3f27addaea-logs\") pod \"3d4332b6-8252-435b-9a1c-bc3f27addaea\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.215002 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-scripts\") pod \"3d4332b6-8252-435b-9a1c-bc3f27addaea\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.215113 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-config-data\") pod \"3d4332b6-8252-435b-9a1c-bc3f27addaea\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.215198 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hjkt\" (UniqueName: \"kubernetes.io/projected/3d4332b6-8252-435b-9a1c-bc3f27addaea-kube-api-access-9hjkt\") pod \"3d4332b6-8252-435b-9a1c-bc3f27addaea\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.215230 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-tls-certs\") pod \"3d4332b6-8252-435b-9a1c-bc3f27addaea\" (UID: \"3d4332b6-8252-435b-9a1c-bc3f27addaea\") " Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.215294 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4332b6-8252-435b-9a1c-bc3f27addaea-logs" (OuterVolumeSpecName: "logs") pod "3d4332b6-8252-435b-9a1c-bc3f27addaea" (UID: "3d4332b6-8252-435b-9a1c-bc3f27addaea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.218506 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d4332b6-8252-435b-9a1c-bc3f27addaea-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.224884 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3d4332b6-8252-435b-9a1c-bc3f27addaea" (UID: "3d4332b6-8252-435b-9a1c-bc3f27addaea"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.236960 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4332b6-8252-435b-9a1c-bc3f27addaea-kube-api-access-9hjkt" (OuterVolumeSpecName: "kube-api-access-9hjkt") pod "3d4332b6-8252-435b-9a1c-bc3f27addaea" (UID: "3d4332b6-8252-435b-9a1c-bc3f27addaea"). InnerVolumeSpecName "kube-api-access-9hjkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.246905 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-config-data" (OuterVolumeSpecName: "config-data") pod "3d4332b6-8252-435b-9a1c-bc3f27addaea" (UID: "3d4332b6-8252-435b-9a1c-bc3f27addaea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.277358 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-scripts" (OuterVolumeSpecName: "scripts") pod "3d4332b6-8252-435b-9a1c-bc3f27addaea" (UID: "3d4332b6-8252-435b-9a1c-bc3f27addaea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.284031 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "3d4332b6-8252-435b-9a1c-bc3f27addaea" (UID: "3d4332b6-8252-435b-9a1c-bc3f27addaea"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.298893 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d4332b6-8252-435b-9a1c-bc3f27addaea" (UID: "3d4332b6-8252-435b-9a1c-bc3f27addaea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.320188 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.320227 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hjkt\" (UniqueName: \"kubernetes.io/projected/3d4332b6-8252-435b-9a1c-bc3f27addaea-kube-api-access-9hjkt\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.320238 4957 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.320247 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.320259 4957 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d4332b6-8252-435b-9a1c-bc3f27addaea-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.320267 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d4332b6-8252-435b-9a1c-bc3f27addaea-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.369302 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerStarted","Data":"d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437"} Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.371306 4957 generic.go:334] "Generic (PLEG): container finished" podID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerID="1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48" exitCode=137 Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.371374 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c6cdbd84-nmkc9" event={"ID":"3d4332b6-8252-435b-9a1c-bc3f27addaea","Type":"ContainerDied","Data":"1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48"} Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.371384 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c6cdbd84-nmkc9" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.371404 4957 scope.go:117] "RemoveContainer" containerID="da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.371392 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c6cdbd84-nmkc9" event={"ID":"3d4332b6-8252-435b-9a1c-bc3f27addaea","Type":"ContainerDied","Data":"f62c753849f8e5f33d057b0da12e2d4917abe7a5d94603dfcced8744ef0f6e41"} Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.380217 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2319a013-4227-4643-9b7f-d60b4d501af7","Type":"ContainerStarted","Data":"ea1ef257715838442502a08bf15547e4b5b69fa24ffe7b4b7606a3d84d7cd82a"} Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.407200 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.40718414 podStartE2EDuration="4.40718414s" podCreationTimestamp="2025-12-06 06:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:45.40680992 +0000 UTC m=+1465.057077552" watchObservedRunningTime="2025-12-06 06:03:45.40718414 +0000 UTC m=+1465.057451772" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.438983 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76c6cdbd84-nmkc9"] Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.455711 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-76c6cdbd84-nmkc9"] Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.532286 4957 scope.go:117] "RemoveContainer" containerID="1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.568335 4957 scope.go:117] "RemoveContainer" containerID="da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b" Dec 06 06:03:45 crc kubenswrapper[4957]: E1206 06:03:45.568948 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b\": container with ID starting with da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b not found: ID does not exist" containerID="da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.568988 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b"} err="failed to get container status \"da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b\": rpc error: code = NotFound desc = could not find container \"da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b\": container with ID starting with da8ae0882dcc0b7ba02d3b30d9c2983748bc9cc174d9dc41a9d774ccbb63906b not found: ID does not exist" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.569013 4957 scope.go:117] "RemoveContainer" containerID="1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48" Dec 06 06:03:45 crc kubenswrapper[4957]: E1206 06:03:45.569200 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48\": container with ID starting with 1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48 not found: ID does not exist" containerID="1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48" Dec 06 06:03:45 crc kubenswrapper[4957]: I1206 06:03:45.569221 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48"} err="failed to get container status \"1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48\": rpc error: code = NotFound desc = could not find container \"1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48\": container with ID starting with 1c599e2fae5e9f4a0fcda455e0c3e50af3ac243a968d457cdc2e93e9569d3b48 not found: ID does not exist" Dec 06 06:03:46 crc kubenswrapper[4957]: I1206 06:03:46.674023 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" path="/var/lib/kubelet/pods/3d4332b6-8252-435b-9a1c-bc3f27addaea/volumes" Dec 06 06:03:47 crc kubenswrapper[4957]: I1206 06:03:47.409086 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerStarted","Data":"b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b"} Dec 06 06:03:49 crc kubenswrapper[4957]: I1206 06:03:49.428058 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerStarted","Data":"15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55"} Dec 06 06:03:49 crc kubenswrapper[4957]: I1206 06:03:49.429498 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:03:49 crc kubenswrapper[4957]: I1206 06:03:49.428251 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="proxy-httpd" containerID="cri-o://15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55" gracePeriod=30 Dec 06 06:03:49 crc kubenswrapper[4957]: I1206 06:03:49.428281 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-notification-agent" containerID="cri-o://d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437" gracePeriod=30 Dec 06 06:03:49 crc kubenswrapper[4957]: I1206 06:03:49.428307 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="sg-core" containerID="cri-o://b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b" gracePeriod=30 Dec 06 06:03:49 crc kubenswrapper[4957]: I1206 06:03:49.428219 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-central-agent" containerID="cri-o://1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4" gracePeriod=30 Dec 06 06:03:49 crc kubenswrapper[4957]: I1206 06:03:49.449005 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.310794015 podStartE2EDuration="8.448984436s" podCreationTimestamp="2025-12-06 06:03:41 +0000 UTC" firstStartedPulling="2025-12-06 06:03:42.371176604 +0000 UTC m=+1462.021444236" lastFinishedPulling="2025-12-06 06:03:48.509367025 +0000 UTC m=+1468.159634657" observedRunningTime="2025-12-06 06:03:49.446730955 +0000 UTC m=+1469.096998587" watchObservedRunningTime="2025-12-06 06:03:49.448984436 +0000 UTC m=+1469.099252068" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.444095 4957 generic.go:334] "Generic (PLEG): container finished" podID="f9f19718-7793-4223-83af-373ffe31a15b" containerID="15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55" exitCode=0 Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.444553 4957 generic.go:334] "Generic (PLEG): container finished" podID="f9f19718-7793-4223-83af-373ffe31a15b" containerID="b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b" exitCode=2 Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.444577 4957 generic.go:334] "Generic (PLEG): container finished" podID="f9f19718-7793-4223-83af-373ffe31a15b" containerID="d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437" exitCode=0 Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.444319 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerDied","Data":"15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55"} Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.444640 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerDied","Data":"b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b"} Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.444668 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerDied","Data":"d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437"} Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.945762 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xfhvr"] Dec 06 06:03:50 crc kubenswrapper[4957]: E1206 06:03:50.946172 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon-log" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.946188 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon-log" Dec 06 06:03:50 crc kubenswrapper[4957]: E1206 06:03:50.946206 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.946212 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.946381 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon-log" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.946414 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4332b6-8252-435b-9a1c-bc3f27addaea" containerName="horizon" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.947037 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.955484 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8e3f-account-create-update-5vz7b"] Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.956877 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.959200 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.976465 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xfhvr"] Dec 06 06:03:50 crc kubenswrapper[4957]: I1206 06:03:50.983460 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8e3f-account-create-update-5vz7b"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.070988 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-4759m"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.072240 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.080544 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4759m"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.128325 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtwbk\" (UniqueName: \"kubernetes.io/projected/c35ff771-af68-49da-91a9-f02b8ffb2d6a-kube-api-access-rtwbk\") pod \"nova-api-db-create-xfhvr\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.129849 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-operator-scripts\") pod \"nova-api-8e3f-account-create-update-5vz7b\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.130098 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35ff771-af68-49da-91a9-f02b8ffb2d6a-operator-scripts\") pod \"nova-api-db-create-xfhvr\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.130386 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59plx\" (UniqueName: \"kubernetes.io/projected/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-kube-api-access-59plx\") pod \"nova-api-8e3f-account-create-update-5vz7b\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.146348 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f97e-account-create-update-vgh4p"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.147649 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.150024 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.157099 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f97e-account-create-update-vgh4p"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232560 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-operator-scripts\") pod \"nova-cell0-f97e-account-create-update-vgh4p\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232636 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35ff771-af68-49da-91a9-f02b8ffb2d6a-operator-scripts\") pod \"nova-api-db-create-xfhvr\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232744 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59plx\" (UniqueName: \"kubernetes.io/projected/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-kube-api-access-59plx\") pod \"nova-api-8e3f-account-create-update-5vz7b\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232787 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tcrm\" (UniqueName: \"kubernetes.io/projected/8aaf6832-b63f-418e-be87-b74fd650c210-kube-api-access-7tcrm\") pod \"nova-cell0-db-create-4759m\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232823 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aaf6832-b63f-418e-be87-b74fd650c210-operator-scripts\") pod \"nova-cell0-db-create-4759m\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232868 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtwbk\" (UniqueName: \"kubernetes.io/projected/c35ff771-af68-49da-91a9-f02b8ffb2d6a-kube-api-access-rtwbk\") pod \"nova-api-db-create-xfhvr\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232934 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-operator-scripts\") pod \"nova-api-8e3f-account-create-update-5vz7b\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.232960 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfhsp\" (UniqueName: \"kubernetes.io/projected/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-kube-api-access-gfhsp\") pod \"nova-cell0-f97e-account-create-update-vgh4p\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.233572 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35ff771-af68-49da-91a9-f02b8ffb2d6a-operator-scripts\") pod \"nova-api-db-create-xfhvr\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.234136 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-operator-scripts\") pod \"nova-api-8e3f-account-create-update-5vz7b\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.248650 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-sv4lc"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.249883 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.264656 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-sv4lc"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.275668 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59plx\" (UniqueName: \"kubernetes.io/projected/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-kube-api-access-59plx\") pod \"nova-api-8e3f-account-create-update-5vz7b\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.287412 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtwbk\" (UniqueName: \"kubernetes.io/projected/c35ff771-af68-49da-91a9-f02b8ffb2d6a-kube-api-access-rtwbk\") pod \"nova-api-db-create-xfhvr\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.324233 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.334541 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tcrm\" (UniqueName: \"kubernetes.io/projected/8aaf6832-b63f-418e-be87-b74fd650c210-kube-api-access-7tcrm\") pod \"nova-cell0-db-create-4759m\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.334618 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aaf6832-b63f-418e-be87-b74fd650c210-operator-scripts\") pod \"nova-cell0-db-create-4759m\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.334681 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfhsp\" (UniqueName: \"kubernetes.io/projected/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-kube-api-access-gfhsp\") pod \"nova-cell0-f97e-account-create-update-vgh4p\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.334719 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-operator-scripts\") pod \"nova-cell0-f97e-account-create-update-vgh4p\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.335376 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-operator-scripts\") pod \"nova-cell0-f97e-account-create-update-vgh4p\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.336110 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aaf6832-b63f-418e-be87-b74fd650c210-operator-scripts\") pod \"nova-cell0-db-create-4759m\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.364122 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdh8p"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.371639 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tcrm\" (UniqueName: \"kubernetes.io/projected/8aaf6832-b63f-418e-be87-b74fd650c210-kube-api-access-7tcrm\") pod \"nova-cell0-db-create-4759m\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.383460 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.388420 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfhsp\" (UniqueName: \"kubernetes.io/projected/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-kube-api-access-gfhsp\") pod \"nova-cell0-f97e-account-create-update-vgh4p\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.392231 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.401321 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdh8p"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.436807 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzb5t\" (UniqueName: \"kubernetes.io/projected/63851a7a-b1a9-4e6e-9260-63d2b9603a16-kube-api-access-dzb5t\") pod \"nova-cell1-db-create-sv4lc\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.436984 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63851a7a-b1a9-4e6e-9260-63d2b9603a16-operator-scripts\") pod \"nova-cell1-db-create-sv4lc\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.452714 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-e1ca-account-create-update-gr9fq"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.454139 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.463069 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.463475 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.475756 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e1ca-account-create-update-gr9fq"] Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.541864 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-utilities\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.541957 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63851a7a-b1a9-4e6e-9260-63d2b9603a16-operator-scripts\") pod \"nova-cell1-db-create-sv4lc\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.542004 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xs4v\" (UniqueName: \"kubernetes.io/projected/33edcf87-18b7-4f62-892d-6ab127062c66-kube-api-access-8xs4v\") pod \"nova-cell1-e1ca-account-create-update-gr9fq\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.542046 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxcq6\" (UniqueName: \"kubernetes.io/projected/a497f65e-6269-468a-99e6-bad35bcb75b8-kube-api-access-hxcq6\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.542072 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-catalog-content\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.542115 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzb5t\" (UniqueName: \"kubernetes.io/projected/63851a7a-b1a9-4e6e-9260-63d2b9603a16-kube-api-access-dzb5t\") pod \"nova-cell1-db-create-sv4lc\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.542175 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33edcf87-18b7-4f62-892d-6ab127062c66-operator-scripts\") pod \"nova-cell1-e1ca-account-create-update-gr9fq\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.542985 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63851a7a-b1a9-4e6e-9260-63d2b9603a16-operator-scripts\") pod \"nova-cell1-db-create-sv4lc\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.562127 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzb5t\" (UniqueName: \"kubernetes.io/projected/63851a7a-b1a9-4e6e-9260-63d2b9603a16-kube-api-access-dzb5t\") pod \"nova-cell1-db-create-sv4lc\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.567568 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.568392 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.643861 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxcq6\" (UniqueName: \"kubernetes.io/projected/a497f65e-6269-468a-99e6-bad35bcb75b8-kube-api-access-hxcq6\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.644131 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-catalog-content\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.644218 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33edcf87-18b7-4f62-892d-6ab127062c66-operator-scripts\") pod \"nova-cell1-e1ca-account-create-update-gr9fq\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.644277 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-utilities\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.644347 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xs4v\" (UniqueName: \"kubernetes.io/projected/33edcf87-18b7-4f62-892d-6ab127062c66-kube-api-access-8xs4v\") pod \"nova-cell1-e1ca-account-create-update-gr9fq\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.644895 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-catalog-content\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.645406 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33edcf87-18b7-4f62-892d-6ab127062c66-operator-scripts\") pod \"nova-cell1-e1ca-account-create-update-gr9fq\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.650001 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-utilities\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.664191 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxcq6\" (UniqueName: \"kubernetes.io/projected/a497f65e-6269-468a-99e6-bad35bcb75b8-kube-api-access-hxcq6\") pod \"redhat-operators-sdh8p\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.684554 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xs4v\" (UniqueName: \"kubernetes.io/projected/33edcf87-18b7-4f62-892d-6ab127062c66-kube-api-access-8xs4v\") pod \"nova-cell1-e1ca-account-create-update-gr9fq\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.757639 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.757722 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.802971 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.834299 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.851462 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.851697 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.878934 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.886730 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.920659 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:03:51 crc kubenswrapper[4957]: I1206 06:03:51.920784 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.023361 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8e3f-account-create-update-5vz7b"] Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.106552 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4759m"] Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.226292 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-sv4lc"] Dec 06 06:03:52 crc kubenswrapper[4957]: W1206 06:03:52.241429 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63851a7a_b1a9_4e6e_9260_63d2b9603a16.slice/crio-ad1caf1b81691353cf87663db62a529f4dc4b8b5a80470dccc3b217f51797cbe WatchSource:0}: Error finding container ad1caf1b81691353cf87663db62a529f4dc4b8b5a80470dccc3b217f51797cbe: Status 404 returned error can't find the container with id ad1caf1b81691353cf87663db62a529f4dc4b8b5a80470dccc3b217f51797cbe Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.420815 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xfhvr"] Dec 06 06:03:52 crc kubenswrapper[4957]: W1206 06:03:52.427305 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5080f24_5973_46b2_a7b0_e5c74a8d7e72.slice/crio-a86b05166f067b700e9501074edc3d16007664dff4bd00708e9b1ca100ff3ca5 WatchSource:0}: Error finding container a86b05166f067b700e9501074edc3d16007664dff4bd00708e9b1ca100ff3ca5: Status 404 returned error can't find the container with id a86b05166f067b700e9501074edc3d16007664dff4bd00708e9b1ca100ff3ca5 Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.442884 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f97e-account-create-update-vgh4p"] Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.478891 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sv4lc" event={"ID":"63851a7a-b1a9-4e6e-9260-63d2b9603a16","Type":"ContainerStarted","Data":"ad1caf1b81691353cf87663db62a529f4dc4b8b5a80470dccc3b217f51797cbe"} Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.481855 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xfhvr" event={"ID":"c35ff771-af68-49da-91a9-f02b8ffb2d6a","Type":"ContainerStarted","Data":"6e2297268f79a12b6056db93a13592162006588dccce7bd4ffdee418f5c4cc4a"} Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.483537 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4759m" event={"ID":"8aaf6832-b63f-418e-be87-b74fd650c210","Type":"ContainerStarted","Data":"4ed5299c98d2a85a0ceb9c0c89b47e45026903b29874b3d9f4616bec83d80088"} Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.484614 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" event={"ID":"b5080f24-5973-46b2-a7b0-e5c74a8d7e72","Type":"ContainerStarted","Data":"a86b05166f067b700e9501074edc3d16007664dff4bd00708e9b1ca100ff3ca5"} Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.489664 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" event={"ID":"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58","Type":"ContainerStarted","Data":"fcdfdd971ae722384e9f68f6649006bcff07db77f172b10fd75bac9228ba6ddd"} Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.489715 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.490914 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.490947 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.490962 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.637638 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdh8p"] Dec 06 06:03:52 crc kubenswrapper[4957]: I1206 06:03:52.651725 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e1ca-account-create-update-gr9fq"] Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.377084 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454286 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-run-httpd\") pod \"f9f19718-7793-4223-83af-373ffe31a15b\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454356 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhbwn\" (UniqueName: \"kubernetes.io/projected/f9f19718-7793-4223-83af-373ffe31a15b-kube-api-access-lhbwn\") pod \"f9f19718-7793-4223-83af-373ffe31a15b\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454402 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-combined-ca-bundle\") pod \"f9f19718-7793-4223-83af-373ffe31a15b\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454456 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-config-data\") pod \"f9f19718-7793-4223-83af-373ffe31a15b\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454491 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-log-httpd\") pod \"f9f19718-7793-4223-83af-373ffe31a15b\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454534 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-scripts\") pod \"f9f19718-7793-4223-83af-373ffe31a15b\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454638 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f9f19718-7793-4223-83af-373ffe31a15b" (UID: "f9f19718-7793-4223-83af-373ffe31a15b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454897 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f9f19718-7793-4223-83af-373ffe31a15b" (UID: "f9f19718-7793-4223-83af-373ffe31a15b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.454985 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-sg-core-conf-yaml\") pod \"f9f19718-7793-4223-83af-373ffe31a15b\" (UID: \"f9f19718-7793-4223-83af-373ffe31a15b\") " Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.455660 4957 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.455681 4957 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9f19718-7793-4223-83af-373ffe31a15b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.464046 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9f19718-7793-4223-83af-373ffe31a15b-kube-api-access-lhbwn" (OuterVolumeSpecName: "kube-api-access-lhbwn") pod "f9f19718-7793-4223-83af-373ffe31a15b" (UID: "f9f19718-7793-4223-83af-373ffe31a15b"). InnerVolumeSpecName "kube-api-access-lhbwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.523848 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-scripts" (OuterVolumeSpecName: "scripts") pod "f9f19718-7793-4223-83af-373ffe31a15b" (UID: "f9f19718-7793-4223-83af-373ffe31a15b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.527035 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f9f19718-7793-4223-83af-373ffe31a15b" (UID: "f9f19718-7793-4223-83af-373ffe31a15b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.539652 4957 generic.go:334] "Generic (PLEG): container finished" podID="8aaf6832-b63f-418e-be87-b74fd650c210" containerID="aa4c3c143e7986c3dcba5a8151acebb153a83551d2983bae7bb477d1b9ae3d7c" exitCode=0 Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.539725 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4759m" event={"ID":"8aaf6832-b63f-418e-be87-b74fd650c210","Type":"ContainerDied","Data":"aa4c3c143e7986c3dcba5a8151acebb153a83551d2983bae7bb477d1b9ae3d7c"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.558114 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" event={"ID":"b5080f24-5973-46b2-a7b0-e5c74a8d7e72","Type":"ContainerStarted","Data":"f00279cba88970410eea431dcfe35cbbb627a0a43442bdad239de02d570d5396"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.559618 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhbwn\" (UniqueName: \"kubernetes.io/projected/f9f19718-7793-4223-83af-373ffe31a15b-kube-api-access-lhbwn\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.559646 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.559661 4957 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.583392 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" podStartSLOduration=2.5833739749999998 podStartE2EDuration="2.583373975s" podCreationTimestamp="2025-12-06 06:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:53.575003839 +0000 UTC m=+1473.225271471" watchObservedRunningTime="2025-12-06 06:03:53.583373975 +0000 UTC m=+1473.233641607" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.588113 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" event={"ID":"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58","Type":"ContainerStarted","Data":"0925acfb1610a9821c6c182618fec73cb46234ef492255c573888698831e3a14"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.608682 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" podStartSLOduration=3.608664509 podStartE2EDuration="3.608664509s" podCreationTimestamp="2025-12-06 06:03:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:53.604115017 +0000 UTC m=+1473.254382669" watchObservedRunningTime="2025-12-06 06:03:53.608664509 +0000 UTC m=+1473.258932141" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.611753 4957 generic.go:334] "Generic (PLEG): container finished" podID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerID="98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86" exitCode=0 Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.613058 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh8p" event={"ID":"a497f65e-6269-468a-99e6-bad35bcb75b8","Type":"ContainerDied","Data":"98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.613151 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh8p" event={"ID":"a497f65e-6269-468a-99e6-bad35bcb75b8","Type":"ContainerStarted","Data":"f0e6bcd5dff005dd8cb38110399764c3571acfdc0ae557f41253f23754674eab"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.615999 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" event={"ID":"33edcf87-18b7-4f62-892d-6ab127062c66","Type":"ContainerStarted","Data":"5a401c43cf809772aecd69ff47de3fa392cb5155bbe03671983f6b34538cc667"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.616037 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" event={"ID":"33edcf87-18b7-4f62-892d-6ab127062c66","Type":"ContainerStarted","Data":"e21097c6deffd2a6dbb4a18e07ba2bcd22ce76c7d77795ad3eac37bfa1c071be"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.655057 4957 generic.go:334] "Generic (PLEG): container finished" podID="63851a7a-b1a9-4e6e-9260-63d2b9603a16" containerID="011bcbc7d489f817ef03d3d8001ff4be7679d919b15529df0f612b8d7837199a" exitCode=0 Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.655439 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sv4lc" event={"ID":"63851a7a-b1a9-4e6e-9260-63d2b9603a16","Type":"ContainerDied","Data":"011bcbc7d489f817ef03d3d8001ff4be7679d919b15529df0f612b8d7837199a"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.676210 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" podStartSLOduration=2.676190417 podStartE2EDuration="2.676190417s" podCreationTimestamp="2025-12-06 06:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:03:53.672408944 +0000 UTC m=+1473.322676576" watchObservedRunningTime="2025-12-06 06:03:53.676190417 +0000 UTC m=+1473.326458049" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.676730 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xfhvr" event={"ID":"c35ff771-af68-49da-91a9-f02b8ffb2d6a","Type":"ContainerStarted","Data":"5e01d251252448379c8f5523a1847a84733826d8dc28c16afd89780ffaa7c2f7"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.682903 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9f19718-7793-4223-83af-373ffe31a15b" (UID: "f9f19718-7793-4223-83af-373ffe31a15b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.698009 4957 generic.go:334] "Generic (PLEG): container finished" podID="f9f19718-7793-4223-83af-373ffe31a15b" containerID="1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4" exitCode=0 Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.699204 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.699512 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerDied","Data":"1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.699606 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9f19718-7793-4223-83af-373ffe31a15b","Type":"ContainerDied","Data":"43897b185c4110875697ff4e028f414a136109785e14b5c97ff5f43993935600"} Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.699686 4957 scope.go:117] "RemoveContainer" containerID="15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.709961 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-config-data" (OuterVolumeSpecName: "config-data") pod "f9f19718-7793-4223-83af-373ffe31a15b" (UID: "f9f19718-7793-4223-83af-373ffe31a15b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.747683 4957 scope.go:117] "RemoveContainer" containerID="b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.763466 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.763498 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f19718-7793-4223-83af-373ffe31a15b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.772922 4957 scope.go:117] "RemoveContainer" containerID="d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.796188 4957 scope.go:117] "RemoveContainer" containerID="1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.816963 4957 scope.go:117] "RemoveContainer" containerID="15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55" Dec 06 06:03:53 crc kubenswrapper[4957]: E1206 06:03:53.817452 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55\": container with ID starting with 15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55 not found: ID does not exist" containerID="15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.817486 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55"} err="failed to get container status \"15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55\": rpc error: code = NotFound desc = could not find container \"15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55\": container with ID starting with 15a86e997030bfe85e54ce831d12ff292911eb270092c67a47f4950debe2cb55 not found: ID does not exist" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.817513 4957 scope.go:117] "RemoveContainer" containerID="b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b" Dec 06 06:03:53 crc kubenswrapper[4957]: E1206 06:03:53.817812 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b\": container with ID starting with b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b not found: ID does not exist" containerID="b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.817910 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b"} err="failed to get container status \"b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b\": rpc error: code = NotFound desc = could not find container \"b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b\": container with ID starting with b810be8b1663e4bab95906ffe8d4776488bc54c136408f9f1f18117d8d6dfa2b not found: ID does not exist" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.817945 4957 scope.go:117] "RemoveContainer" containerID="d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437" Dec 06 06:03:53 crc kubenswrapper[4957]: E1206 06:03:53.818356 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437\": container with ID starting with d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437 not found: ID does not exist" containerID="d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.818380 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437"} err="failed to get container status \"d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437\": rpc error: code = NotFound desc = could not find container \"d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437\": container with ID starting with d4e2a460fe2caeb0c012977fad5c15a8a2ad9b54447f85a4f8cd94b5c1f9b437 not found: ID does not exist" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.818395 4957 scope.go:117] "RemoveContainer" containerID="1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4" Dec 06 06:03:53 crc kubenswrapper[4957]: E1206 06:03:53.818673 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4\": container with ID starting with 1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4 not found: ID does not exist" containerID="1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4" Dec 06 06:03:53 crc kubenswrapper[4957]: I1206 06:03:53.818698 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4"} err="failed to get container status \"1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4\": rpc error: code = NotFound desc = could not find container \"1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4\": container with ID starting with 1d844cc7cc32b67a3fd104faa2d12afff40e86849692c6d50c00394ab14a5bf4 not found: ID does not exist" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.111641 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.131542 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.142630 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:54 crc kubenswrapper[4957]: E1206 06:03:54.143037 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-notification-agent" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143060 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-notification-agent" Dec 06 06:03:54 crc kubenswrapper[4957]: E1206 06:03:54.143083 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="proxy-httpd" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143089 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="proxy-httpd" Dec 06 06:03:54 crc kubenswrapper[4957]: E1206 06:03:54.143121 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="sg-core" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143126 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="sg-core" Dec 06 06:03:54 crc kubenswrapper[4957]: E1206 06:03:54.143141 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-central-agent" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143150 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-central-agent" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143326 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-notification-agent" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143352 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="sg-core" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143360 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="ceilometer-central-agent" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.143375 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f19718-7793-4223-83af-373ffe31a15b" containerName="proxy-httpd" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.145087 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.151043 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.151626 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.151822 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.272793 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-log-httpd\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.272916 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.273020 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-scripts\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.273145 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-run-httpd\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.273183 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.273213 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7g2d\" (UniqueName: \"kubernetes.io/projected/dd5af873-8d3d-4d80-ab9f-b5003328b687-kube-api-access-w7g2d\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.273306 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.374628 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-log-httpd\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.374701 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.374768 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-scripts\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.374850 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-run-httpd\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.374880 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.374903 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7g2d\" (UniqueName: \"kubernetes.io/projected/dd5af873-8d3d-4d80-ab9f-b5003328b687-kube-api-access-w7g2d\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.374944 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.375270 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-log-httpd\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.375330 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-run-httpd\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.381270 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-scripts\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.381703 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.382333 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.395790 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.401898 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7g2d\" (UniqueName: \"kubernetes.io/projected/dd5af873-8d3d-4d80-ab9f-b5003328b687-kube-api-access-w7g2d\") pod \"ceilometer-0\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.466876 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.693015 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9f19718-7793-4223-83af-373ffe31a15b" path="/var/lib/kubelet/pods/f9f19718-7793-4223-83af-373ffe31a15b/volumes" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.709986 4957 generic.go:334] "Generic (PLEG): container finished" podID="c35ff771-af68-49da-91a9-f02b8ffb2d6a" containerID="5e01d251252448379c8f5523a1847a84733826d8dc28c16afd89780ffaa7c2f7" exitCode=0 Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.710121 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xfhvr" event={"ID":"c35ff771-af68-49da-91a9-f02b8ffb2d6a","Type":"ContainerDied","Data":"5e01d251252448379c8f5523a1847a84733826d8dc28c16afd89780ffaa7c2f7"} Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.721136 4957 generic.go:334] "Generic (PLEG): container finished" podID="b5080f24-5973-46b2-a7b0-e5c74a8d7e72" containerID="f00279cba88970410eea431dcfe35cbbb627a0a43442bdad239de02d570d5396" exitCode=0 Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.721222 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" event={"ID":"b5080f24-5973-46b2-a7b0-e5c74a8d7e72","Type":"ContainerDied","Data":"f00279cba88970410eea431dcfe35cbbb627a0a43442bdad239de02d570d5396"} Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.726952 4957 generic.go:334] "Generic (PLEG): container finished" podID="8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58" containerID="0925acfb1610a9821c6c182618fec73cb46234ef492255c573888698831e3a14" exitCode=0 Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.727041 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" event={"ID":"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58","Type":"ContainerDied","Data":"0925acfb1610a9821c6c182618fec73cb46234ef492255c573888698831e3a14"} Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.731495 4957 generic.go:334] "Generic (PLEG): container finished" podID="33edcf87-18b7-4f62-892d-6ab127062c66" containerID="5a401c43cf809772aecd69ff47de3fa392cb5155bbe03671983f6b34538cc667" exitCode=0 Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.731586 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.731594 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.732406 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" event={"ID":"33edcf87-18b7-4f62-892d-6ab127062c66","Type":"ContainerDied","Data":"5a401c43cf809772aecd69ff47de3fa392cb5155bbe03671983f6b34538cc667"} Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.732535 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.732544 4957 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 06:03:54 crc kubenswrapper[4957]: I1206 06:03:54.960185 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.020416 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.089827 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.092374 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.189057 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tcrm\" (UniqueName: \"kubernetes.io/projected/8aaf6832-b63f-418e-be87-b74fd650c210-kube-api-access-7tcrm\") pod \"8aaf6832-b63f-418e-be87-b74fd650c210\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.189150 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aaf6832-b63f-418e-be87-b74fd650c210-operator-scripts\") pod \"8aaf6832-b63f-418e-be87-b74fd650c210\" (UID: \"8aaf6832-b63f-418e-be87-b74fd650c210\") " Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.192234 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aaf6832-b63f-418e-be87-b74fd650c210-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8aaf6832-b63f-418e-be87-b74fd650c210" (UID: "8aaf6832-b63f-418e-be87-b74fd650c210"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.202060 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aaf6832-b63f-418e-be87-b74fd650c210-kube-api-access-7tcrm" (OuterVolumeSpecName: "kube-api-access-7tcrm") pod "8aaf6832-b63f-418e-be87-b74fd650c210" (UID: "8aaf6832-b63f-418e-be87-b74fd650c210"). InnerVolumeSpecName "kube-api-access-7tcrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.250790 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.257409 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.274083 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.293196 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tcrm\" (UniqueName: \"kubernetes.io/projected/8aaf6832-b63f-418e-be87-b74fd650c210-kube-api-access-7tcrm\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.293471 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8aaf6832-b63f-418e-be87-b74fd650c210-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.394173 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63851a7a-b1a9-4e6e-9260-63d2b9603a16-operator-scripts\") pod \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.394218 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35ff771-af68-49da-91a9-f02b8ffb2d6a-operator-scripts\") pod \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.394308 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtwbk\" (UniqueName: \"kubernetes.io/projected/c35ff771-af68-49da-91a9-f02b8ffb2d6a-kube-api-access-rtwbk\") pod \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\" (UID: \"c35ff771-af68-49da-91a9-f02b8ffb2d6a\") " Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.394358 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzb5t\" (UniqueName: \"kubernetes.io/projected/63851a7a-b1a9-4e6e-9260-63d2b9603a16-kube-api-access-dzb5t\") pod \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\" (UID: \"63851a7a-b1a9-4e6e-9260-63d2b9603a16\") " Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.395478 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63851a7a-b1a9-4e6e-9260-63d2b9603a16-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "63851a7a-b1a9-4e6e-9260-63d2b9603a16" (UID: "63851a7a-b1a9-4e6e-9260-63d2b9603a16"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.395845 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35ff771-af68-49da-91a9-f02b8ffb2d6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c35ff771-af68-49da-91a9-f02b8ffb2d6a" (UID: "c35ff771-af68-49da-91a9-f02b8ffb2d6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.408049 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35ff771-af68-49da-91a9-f02b8ffb2d6a-kube-api-access-rtwbk" (OuterVolumeSpecName: "kube-api-access-rtwbk") pod "c35ff771-af68-49da-91a9-f02b8ffb2d6a" (UID: "c35ff771-af68-49da-91a9-f02b8ffb2d6a"). InnerVolumeSpecName "kube-api-access-rtwbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.408113 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63851a7a-b1a9-4e6e-9260-63d2b9603a16-kube-api-access-dzb5t" (OuterVolumeSpecName: "kube-api-access-dzb5t") pod "63851a7a-b1a9-4e6e-9260-63d2b9603a16" (UID: "63851a7a-b1a9-4e6e-9260-63d2b9603a16"). InnerVolumeSpecName "kube-api-access-dzb5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.409082 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.498894 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63851a7a-b1a9-4e6e-9260-63d2b9603a16-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.499195 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c35ff771-af68-49da-91a9-f02b8ffb2d6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.499207 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtwbk\" (UniqueName: \"kubernetes.io/projected/c35ff771-af68-49da-91a9-f02b8ffb2d6a-kube-api-access-rtwbk\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.499218 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzb5t\" (UniqueName: \"kubernetes.io/projected/63851a7a-b1a9-4e6e-9260-63d2b9603a16-kube-api-access-dzb5t\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.744642 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4759m" event={"ID":"8aaf6832-b63f-418e-be87-b74fd650c210","Type":"ContainerDied","Data":"4ed5299c98d2a85a0ceb9c0c89b47e45026903b29874b3d9f4616bec83d80088"} Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.744697 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ed5299c98d2a85a0ceb9c0c89b47e45026903b29874b3d9f4616bec83d80088" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.744933 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4759m" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.750675 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerStarted","Data":"5d4a075b7963ba13d110fe4bb977ac59d366998134fa58c98ad2c8f6a73e2675"} Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.753144 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sv4lc" event={"ID":"63851a7a-b1a9-4e6e-9260-63d2b9603a16","Type":"ContainerDied","Data":"ad1caf1b81691353cf87663db62a529f4dc4b8b5a80470dccc3b217f51797cbe"} Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.753370 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad1caf1b81691353cf87663db62a529f4dc4b8b5a80470dccc3b217f51797cbe" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.753181 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sv4lc" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.755631 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xfhvr" Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.755659 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xfhvr" event={"ID":"c35ff771-af68-49da-91a9-f02b8ffb2d6a","Type":"ContainerDied","Data":"6e2297268f79a12b6056db93a13592162006588dccce7bd4ffdee418f5c4cc4a"} Dec 06 06:03:55 crc kubenswrapper[4957]: I1206 06:03:55.756198 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e2297268f79a12b6056db93a13592162006588dccce7bd4ffdee418f5c4cc4a" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.180077 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.203496 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.327952 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfhsp\" (UniqueName: \"kubernetes.io/projected/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-kube-api-access-gfhsp\") pod \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.328038 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-operator-scripts\") pod \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\" (UID: \"b5080f24-5973-46b2-a7b0-e5c74a8d7e72\") " Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.328103 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59plx\" (UniqueName: \"kubernetes.io/projected/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-kube-api-access-59plx\") pod \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.328120 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-operator-scripts\") pod \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\" (UID: \"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58\") " Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.329369 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58" (UID: "8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.330578 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5080f24-5973-46b2-a7b0-e5c74a8d7e72" (UID: "b5080f24-5973-46b2-a7b0-e5c74a8d7e72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.334815 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-kube-api-access-gfhsp" (OuterVolumeSpecName: "kube-api-access-gfhsp") pod "b5080f24-5973-46b2-a7b0-e5c74a8d7e72" (UID: "b5080f24-5973-46b2-a7b0-e5c74a8d7e72"). InnerVolumeSpecName "kube-api-access-gfhsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.337653 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-kube-api-access-59plx" (OuterVolumeSpecName: "kube-api-access-59plx") pod "8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58" (UID: "8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58"). InnerVolumeSpecName "kube-api-access-59plx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.430760 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59plx\" (UniqueName: \"kubernetes.io/projected/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-kube-api-access-59plx\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.430815 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.430850 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfhsp\" (UniqueName: \"kubernetes.io/projected/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-kube-api-access-gfhsp\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.430864 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5080f24-5973-46b2-a7b0-e5c74a8d7e72-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.431802 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.532316 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33edcf87-18b7-4f62-892d-6ab127062c66-operator-scripts\") pod \"33edcf87-18b7-4f62-892d-6ab127062c66\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.532540 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xs4v\" (UniqueName: \"kubernetes.io/projected/33edcf87-18b7-4f62-892d-6ab127062c66-kube-api-access-8xs4v\") pod \"33edcf87-18b7-4f62-892d-6ab127062c66\" (UID: \"33edcf87-18b7-4f62-892d-6ab127062c66\") " Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.533054 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33edcf87-18b7-4f62-892d-6ab127062c66-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "33edcf87-18b7-4f62-892d-6ab127062c66" (UID: "33edcf87-18b7-4f62-892d-6ab127062c66"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.549083 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33edcf87-18b7-4f62-892d-6ab127062c66-kube-api-access-8xs4v" (OuterVolumeSpecName: "kube-api-access-8xs4v") pod "33edcf87-18b7-4f62-892d-6ab127062c66" (UID: "33edcf87-18b7-4f62-892d-6ab127062c66"). InnerVolumeSpecName "kube-api-access-8xs4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.634696 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xs4v\" (UniqueName: \"kubernetes.io/projected/33edcf87-18b7-4f62-892d-6ab127062c66-kube-api-access-8xs4v\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.634733 4957 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33edcf87-18b7-4f62-892d-6ab127062c66-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.767072 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.767124 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f97e-account-create-update-vgh4p" event={"ID":"b5080f24-5973-46b2-a7b0-e5c74a8d7e72","Type":"ContainerDied","Data":"a86b05166f067b700e9501074edc3d16007664dff4bd00708e9b1ca100ff3ca5"} Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.768666 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a86b05166f067b700e9501074edc3d16007664dff4bd00708e9b1ca100ff3ca5" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.772056 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" event={"ID":"8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58","Type":"ContainerDied","Data":"fcdfdd971ae722384e9f68f6649006bcff07db77f172b10fd75bac9228ba6ddd"} Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.772182 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcdfdd971ae722384e9f68f6649006bcff07db77f172b10fd75bac9228ba6ddd" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.772311 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e3f-account-create-update-5vz7b" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.776355 4957 generic.go:334] "Generic (PLEG): container finished" podID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerID="de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea" exitCode=0 Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.776459 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh8p" event={"ID":"a497f65e-6269-468a-99e6-bad35bcb75b8","Type":"ContainerDied","Data":"de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea"} Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.780938 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.781459 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e1ca-account-create-update-gr9fq" event={"ID":"33edcf87-18b7-4f62-892d-6ab127062c66","Type":"ContainerDied","Data":"e21097c6deffd2a6dbb4a18e07ba2bcd22ce76c7d77795ad3eac37bfa1c071be"} Dec 06 06:03:56 crc kubenswrapper[4957]: I1206 06:03:56.781568 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e21097c6deffd2a6dbb4a18e07ba2bcd22ce76c7d77795ad3eac37bfa1c071be" Dec 06 06:04:00 crc kubenswrapper[4957]: I1206 06:04:00.816499 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerStarted","Data":"86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94"} Dec 06 06:04:00 crc kubenswrapper[4957]: I1206 06:04:00.816897 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerStarted","Data":"afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d"} Dec 06 06:04:00 crc kubenswrapper[4957]: I1206 06:04:00.818490 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh8p" event={"ID":"a497f65e-6269-468a-99e6-bad35bcb75b8","Type":"ContainerStarted","Data":"e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5"} Dec 06 06:04:00 crc kubenswrapper[4957]: I1206 06:04:00.836286 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdh8p" podStartSLOduration=3.478823792 podStartE2EDuration="9.836269673s" podCreationTimestamp="2025-12-06 06:03:51 +0000 UTC" firstStartedPulling="2025-12-06 06:03:53.614134727 +0000 UTC m=+1473.264402359" lastFinishedPulling="2025-12-06 06:03:59.971580608 +0000 UTC m=+1479.621848240" observedRunningTime="2025-12-06 06:04:00.83326754 +0000 UTC m=+1480.483535172" watchObservedRunningTime="2025-12-06 06:04:00.836269673 +0000 UTC m=+1480.486537305" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.400241 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ccb8p"] Dec 06 06:04:01 crc kubenswrapper[4957]: E1206 06:04:01.401016 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35ff771-af68-49da-91a9-f02b8ffb2d6a" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401035 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35ff771-af68-49da-91a9-f02b8ffb2d6a" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: E1206 06:04:01.401052 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401059 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: E1206 06:04:01.401071 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aaf6832-b63f-418e-be87-b74fd650c210" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401078 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aaf6832-b63f-418e-be87-b74fd650c210" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: E1206 06:04:01.401095 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33edcf87-18b7-4f62-892d-6ab127062c66" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401100 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="33edcf87-18b7-4f62-892d-6ab127062c66" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: E1206 06:04:01.401111 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63851a7a-b1a9-4e6e-9260-63d2b9603a16" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401117 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="63851a7a-b1a9-4e6e-9260-63d2b9603a16" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: E1206 06:04:01.401136 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5080f24-5973-46b2-a7b0-e5c74a8d7e72" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401143 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5080f24-5973-46b2-a7b0-e5c74a8d7e72" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401306 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="63851a7a-b1a9-4e6e-9260-63d2b9603a16" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401320 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5080f24-5973-46b2-a7b0-e5c74a8d7e72" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401333 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401344 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="33edcf87-18b7-4f62-892d-6ab127062c66" containerName="mariadb-account-create-update" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401355 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aaf6832-b63f-418e-be87-b74fd650c210" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401366 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35ff771-af68-49da-91a9-f02b8ffb2d6a" containerName="mariadb-database-create" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.401967 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.405791 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.406047 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.406241 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wz64c" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.444596 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ccb8p"] Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.545655 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-config-data\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.545718 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w452r\" (UniqueName: \"kubernetes.io/projected/e0cdc965-475d-4d47-b875-9ade77089ae6-kube-api-access-w452r\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.545751 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-scripts\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.545770 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.647400 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-config-data\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.647491 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w452r\" (UniqueName: \"kubernetes.io/projected/e0cdc965-475d-4d47-b875-9ade77089ae6-kube-api-access-w452r\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.647540 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-scripts\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.647563 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.652390 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-config-data\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.652541 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.656404 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-scripts\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.680530 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w452r\" (UniqueName: \"kubernetes.io/projected/e0cdc965-475d-4d47-b875-9ade77089ae6-kube-api-access-w452r\") pod \"nova-cell0-conductor-db-sync-ccb8p\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.800197 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.839509 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerStarted","Data":"3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241"} Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.880245 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:04:01 crc kubenswrapper[4957]: I1206 06:04:01.880286 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:04:02 crc kubenswrapper[4957]: I1206 06:04:02.471100 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ccb8p"] Dec 06 06:04:02 crc kubenswrapper[4957]: W1206 06:04:02.475525 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0cdc965_475d_4d47_b875_9ade77089ae6.slice/crio-0f2c4e4d68692e2ef70694bae199672a1d71514ee398d7994624349a79864c99 WatchSource:0}: Error finding container 0f2c4e4d68692e2ef70694bae199672a1d71514ee398d7994624349a79864c99: Status 404 returned error can't find the container with id 0f2c4e4d68692e2ef70694bae199672a1d71514ee398d7994624349a79864c99 Dec 06 06:04:02 crc kubenswrapper[4957]: I1206 06:04:02.848648 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" event={"ID":"e0cdc965-475d-4d47-b875-9ade77089ae6","Type":"ContainerStarted","Data":"0f2c4e4d68692e2ef70694bae199672a1d71514ee398d7994624349a79864c99"} Dec 06 06:04:02 crc kubenswrapper[4957]: I1206 06:04:02.976487 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sdh8p" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="registry-server" probeResult="failure" output=< Dec 06 06:04:02 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 06:04:02 crc kubenswrapper[4957]: > Dec 06 06:04:03 crc kubenswrapper[4957]: I1206 06:04:03.871945 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerStarted","Data":"57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd"} Dec 06 06:04:03 crc kubenswrapper[4957]: I1206 06:04:03.873520 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:04:03 crc kubenswrapper[4957]: I1206 06:04:03.898603 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.261735799 podStartE2EDuration="9.898586449s" podCreationTimestamp="2025-12-06 06:03:54 +0000 UTC" firstStartedPulling="2025-12-06 06:03:55.12104103 +0000 UTC m=+1474.771308662" lastFinishedPulling="2025-12-06 06:04:02.75789168 +0000 UTC m=+1482.408159312" observedRunningTime="2025-12-06 06:04:03.897312474 +0000 UTC m=+1483.547580106" watchObservedRunningTime="2025-12-06 06:04:03.898586449 +0000 UTC m=+1483.548854081" Dec 06 06:04:05 crc kubenswrapper[4957]: I1206 06:04:05.820897 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fjk65"] Dec 06 06:04:05 crc kubenswrapper[4957]: I1206 06:04:05.823034 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:05 crc kubenswrapper[4957]: I1206 06:04:05.843321 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fjk65"] Dec 06 06:04:05 crc kubenswrapper[4957]: I1206 06:04:05.943028 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mftc\" (UniqueName: \"kubernetes.io/projected/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-kube-api-access-6mftc\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:05 crc kubenswrapper[4957]: I1206 06:04:05.943104 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-catalog-content\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:05 crc kubenswrapper[4957]: I1206 06:04:05.943243 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-utilities\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:06 crc kubenswrapper[4957]: I1206 06:04:06.044586 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mftc\" (UniqueName: \"kubernetes.io/projected/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-kube-api-access-6mftc\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:06 crc kubenswrapper[4957]: I1206 06:04:06.044690 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-catalog-content\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:06 crc kubenswrapper[4957]: I1206 06:04:06.044743 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-utilities\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:06 crc kubenswrapper[4957]: I1206 06:04:06.045955 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-catalog-content\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:06 crc kubenswrapper[4957]: I1206 06:04:06.046333 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-utilities\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:06 crc kubenswrapper[4957]: I1206 06:04:06.063536 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mftc\" (UniqueName: \"kubernetes.io/projected/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-kube-api-access-6mftc\") pod \"community-operators-fjk65\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:06 crc kubenswrapper[4957]: I1206 06:04:06.174858 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.607634 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.608772 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="proxy-httpd" containerID="cri-o://57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd" gracePeriod=30 Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.608946 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="sg-core" containerID="cri-o://3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241" gracePeriod=30 Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.609003 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-notification-agent" containerID="cri-o://86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94" gracePeriod=30 Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.609079 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-central-agent" containerID="cri-o://afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d" gracePeriod=30 Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.998565 4957 generic.go:334] "Generic (PLEG): container finished" podID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerID="57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd" exitCode=0 Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.999344 4957 generic.go:334] "Generic (PLEG): container finished" podID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerID="3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241" exitCode=2 Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.998895 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerDied","Data":"57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd"} Dec 06 06:04:11 crc kubenswrapper[4957]: I1206 06:04:11.999391 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerDied","Data":"3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241"} Dec 06 06:04:12 crc kubenswrapper[4957]: I1206 06:04:12.338364 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fjk65"] Dec 06 06:04:12 crc kubenswrapper[4957]: I1206 06:04:12.938506 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sdh8p" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="registry-server" probeResult="failure" output=< Dec 06 06:04:12 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 06:04:12 crc kubenswrapper[4957]: > Dec 06 06:04:13 crc kubenswrapper[4957]: I1206 06:04:13.011471 4957 generic.go:334] "Generic (PLEG): container finished" podID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerID="afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d" exitCode=0 Dec 06 06:04:13 crc kubenswrapper[4957]: I1206 06:04:13.011530 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerDied","Data":"afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d"} Dec 06 06:04:13 crc kubenswrapper[4957]: I1206 06:04:13.013309 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" event={"ID":"e0cdc965-475d-4d47-b875-9ade77089ae6","Type":"ContainerStarted","Data":"4f2be6a8216ab58840f826fe9a624980c002648bb8b1059ae92ce567ba4c7cfd"} Dec 06 06:04:13 crc kubenswrapper[4957]: I1206 06:04:13.016904 4957 generic.go:334] "Generic (PLEG): container finished" podID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerID="031dc5b0afdd9c48a1f6c8fcbda785ba07bec4be442e51587758cbaa02250e2a" exitCode=0 Dec 06 06:04:13 crc kubenswrapper[4957]: I1206 06:04:13.016941 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjk65" event={"ID":"52576ee1-8b4e-4ec9-906d-8c90e0209a7c","Type":"ContainerDied","Data":"031dc5b0afdd9c48a1f6c8fcbda785ba07bec4be442e51587758cbaa02250e2a"} Dec 06 06:04:13 crc kubenswrapper[4957]: I1206 06:04:13.016962 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjk65" event={"ID":"52576ee1-8b4e-4ec9-906d-8c90e0209a7c","Type":"ContainerStarted","Data":"83215b9c5517f690fd3b16c9dd0886b64102ef7dfbff40eee6e5d9721674cee7"} Dec 06 06:04:13 crc kubenswrapper[4957]: I1206 06:04:13.057317 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" podStartSLOduration=2.710688696 podStartE2EDuration="12.057299796s" podCreationTimestamp="2025-12-06 06:04:01 +0000 UTC" firstStartedPulling="2025-12-06 06:04:02.477476826 +0000 UTC m=+1482.127744468" lastFinishedPulling="2025-12-06 06:04:11.824087936 +0000 UTC m=+1491.474355568" observedRunningTime="2025-12-06 06:04:13.050340478 +0000 UTC m=+1492.700608110" watchObservedRunningTime="2025-12-06 06:04:13.057299796 +0000 UTC m=+1492.707567428" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.017493 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.042274 4957 generic.go:334] "Generic (PLEG): container finished" podID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerID="86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94" exitCode=0 Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.042337 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerDied","Data":"86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94"} Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.042367 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd5af873-8d3d-4d80-ab9f-b5003328b687","Type":"ContainerDied","Data":"5d4a075b7963ba13d110fe4bb977ac59d366998134fa58c98ad2c8f6a73e2675"} Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.042389 4957 scope.go:117] "RemoveContainer" containerID="57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.042520 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.049953 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjk65" event={"ID":"52576ee1-8b4e-4ec9-906d-8c90e0209a7c","Type":"ContainerStarted","Data":"a6df5585144e16ecba6e33db4c08823fad841a2564c557c5a8694006387bbbf8"} Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.078404 4957 scope.go:117] "RemoveContainer" containerID="3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.103156 4957 scope.go:117] "RemoveContainer" containerID="86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.126380 4957 scope.go:117] "RemoveContainer" containerID="afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134129 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134197 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-scripts\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134307 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-run-httpd\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134365 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-sg-core-conf-yaml\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134423 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-log-httpd\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134548 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7g2d\" (UniqueName: \"kubernetes.io/projected/dd5af873-8d3d-4d80-ab9f-b5003328b687-kube-api-access-w7g2d\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134598 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-combined-ca-bundle\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134741 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.134963 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.135407 4957 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.135424 4957 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd5af873-8d3d-4d80-ab9f-b5003328b687-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.139397 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd5af873-8d3d-4d80-ab9f-b5003328b687-kube-api-access-w7g2d" (OuterVolumeSpecName: "kube-api-access-w7g2d") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "kube-api-access-w7g2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.159770 4957 scope.go:117] "RemoveContainer" containerID="57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.160033 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-scripts" (OuterVolumeSpecName: "scripts") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.160452 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd\": container with ID starting with 57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd not found: ID does not exist" containerID="57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.160488 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd"} err="failed to get container status \"57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd\": rpc error: code = NotFound desc = could not find container \"57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd\": container with ID starting with 57a984b6c29d0800b2bfef94ba1c44a3526b62a53d9c31a71eb92a2dbe1547bd not found: ID does not exist" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.160513 4957 scope.go:117] "RemoveContainer" containerID="3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241" Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.161961 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241\": container with ID starting with 3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241 not found: ID does not exist" containerID="3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.162002 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241"} err="failed to get container status \"3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241\": rpc error: code = NotFound desc = could not find container \"3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241\": container with ID starting with 3316aaa80d8493dfef014682d5c928886e67d4ef0c96a8896e34d22a0bae0241 not found: ID does not exist" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.162184 4957 scope.go:117] "RemoveContainer" containerID="86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94" Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.163237 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94\": container with ID starting with 86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94 not found: ID does not exist" containerID="86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.163323 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94"} err="failed to get container status \"86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94\": rpc error: code = NotFound desc = could not find container \"86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94\": container with ID starting with 86c9b0151033bf33633b4fb4f94b5d9aa42cc36b1fc3d91ab0e18c7ab7fb2b94 not found: ID does not exist" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.163392 4957 scope.go:117] "RemoveContainer" containerID="afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d" Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.163966 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d\": container with ID starting with afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d not found: ID does not exist" containerID="afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.163997 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d"} err="failed to get container status \"afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d\": rpc error: code = NotFound desc = could not find container \"afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d\": container with ID starting with afb146bc18f14d599bfd061c932c20612c8ff17d8c9aa45e9bd9ae756e09ad4d not found: ID does not exist" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.191339 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.232961 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.236070 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data" (OuterVolumeSpecName: "config-data") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.237039 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data\") pod \"dd5af873-8d3d-4d80-ab9f-b5003328b687\" (UID: \"dd5af873-8d3d-4d80-ab9f-b5003328b687\") " Dec 06 06:04:15 crc kubenswrapper[4957]: W1206 06:04:15.237277 4957 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/dd5af873-8d3d-4d80-ab9f-b5003328b687/volumes/kubernetes.io~secret/config-data Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.237296 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data" (OuterVolumeSpecName: "config-data") pod "dd5af873-8d3d-4d80-ab9f-b5003328b687" (UID: "dd5af873-8d3d-4d80-ab9f-b5003328b687"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.237667 4957 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.237693 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7g2d\" (UniqueName: \"kubernetes.io/projected/dd5af873-8d3d-4d80-ab9f-b5003328b687-kube-api-access-w7g2d\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.237708 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.237784 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.237796 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd5af873-8d3d-4d80-ab9f-b5003328b687-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.379973 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.392881 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.407406 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.407899 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-notification-agent" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.407926 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-notification-agent" Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.407946 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-central-agent" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.407956 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-central-agent" Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.407980 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="proxy-httpd" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.407989 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="proxy-httpd" Dec 06 06:04:15 crc kubenswrapper[4957]: E1206 06:04:15.408028 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="sg-core" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.408039 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="sg-core" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.408262 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-notification-agent" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.408283 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="sg-core" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.408302 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="ceilometer-central-agent" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.408316 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" containerName="proxy-httpd" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.410713 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.416196 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.416419 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.436442 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.542776 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-config-data\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.542858 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-scripts\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.542895 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.542925 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.543114 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-run-httpd\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.543495 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44fz6\" (UniqueName: \"kubernetes.io/projected/97341258-bed0-4dda-9b61-bf6e9c93d375-kube-api-access-44fz6\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.543526 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-log-httpd\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.645544 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44fz6\" (UniqueName: \"kubernetes.io/projected/97341258-bed0-4dda-9b61-bf6e9c93d375-kube-api-access-44fz6\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.645586 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-log-httpd\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.645670 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-config-data\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.645695 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-scripts\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.645719 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.645747 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.645778 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-run-httpd\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.646680 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-log-httpd\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.646705 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-run-httpd\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.649443 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.649537 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-scripts\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.650351 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.650582 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-config-data\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.661928 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44fz6\" (UniqueName: \"kubernetes.io/projected/97341258-bed0-4dda-9b61-bf6e9c93d375-kube-api-access-44fz6\") pod \"ceilometer-0\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " pod="openstack/ceilometer-0" Dec 06 06:04:15 crc kubenswrapper[4957]: I1206 06:04:15.731247 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:16 crc kubenswrapper[4957]: I1206 06:04:16.200918 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:16 crc kubenswrapper[4957]: W1206 06:04:16.207439 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97341258_bed0_4dda_9b61_bf6e9c93d375.slice/crio-cbb0c6fa736e2dba51275abf9b3fd30b0e4b428807c8bb5a4c12f3ebebac7b34 WatchSource:0}: Error finding container cbb0c6fa736e2dba51275abf9b3fd30b0e4b428807c8bb5a4c12f3ebebac7b34: Status 404 returned error can't find the container with id cbb0c6fa736e2dba51275abf9b3fd30b0e4b428807c8bb5a4c12f3ebebac7b34 Dec 06 06:04:16 crc kubenswrapper[4957]: I1206 06:04:16.673893 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd5af873-8d3d-4d80-ab9f-b5003328b687" path="/var/lib/kubelet/pods/dd5af873-8d3d-4d80-ab9f-b5003328b687/volumes" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.072911 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerStarted","Data":"cbb0c6fa736e2dba51275abf9b3fd30b0e4b428807c8bb5a4c12f3ebebac7b34"} Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.074867 4957 generic.go:334] "Generic (PLEG): container finished" podID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerID="a6df5585144e16ecba6e33db4c08823fad841a2564c557c5a8694006387bbbf8" exitCode=0 Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.074925 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjk65" event={"ID":"52576ee1-8b4e-4ec9-906d-8c90e0209a7c","Type":"ContainerDied","Data":"a6df5585144e16ecba6e33db4c08823fad841a2564c557c5a8694006387bbbf8"} Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.602621 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4cp5j"] Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.606275 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.613354 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cp5j"] Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.779004 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-utilities\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.780002 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-catalog-content\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.780277 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r2pf\" (UniqueName: \"kubernetes.io/projected/14414c4f-7e4f-4dfb-b44c-4a3427c36675-kube-api-access-5r2pf\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.882153 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r2pf\" (UniqueName: \"kubernetes.io/projected/14414c4f-7e4f-4dfb-b44c-4a3427c36675-kube-api-access-5r2pf\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.882234 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-utilities\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.882307 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-catalog-content\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.882758 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-utilities\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.882792 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-catalog-content\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.907905 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r2pf\" (UniqueName: \"kubernetes.io/projected/14414c4f-7e4f-4dfb-b44c-4a3427c36675-kube-api-access-5r2pf\") pod \"redhat-marketplace-4cp5j\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:17 crc kubenswrapper[4957]: I1206 06:04:17.926244 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:18 crc kubenswrapper[4957]: I1206 06:04:18.993191 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cp5j"] Dec 06 06:04:18 crc kubenswrapper[4957]: W1206 06:04:18.995664 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14414c4f_7e4f_4dfb_b44c_4a3427c36675.slice/crio-5aa67941d447871d63900142843c6fa7415ec2ed64d49b83b97d49ff650c5d5a WatchSource:0}: Error finding container 5aa67941d447871d63900142843c6fa7415ec2ed64d49b83b97d49ff650c5d5a: Status 404 returned error can't find the container with id 5aa67941d447871d63900142843c6fa7415ec2ed64d49b83b97d49ff650c5d5a Dec 06 06:04:19 crc kubenswrapper[4957]: I1206 06:04:19.092859 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerStarted","Data":"fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a"} Dec 06 06:04:19 crc kubenswrapper[4957]: I1206 06:04:19.094477 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerStarted","Data":"5aa67941d447871d63900142843c6fa7415ec2ed64d49b83b97d49ff650c5d5a"} Dec 06 06:04:21 crc kubenswrapper[4957]: I1206 06:04:21.113637 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerStarted","Data":"c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e"} Dec 06 06:04:21 crc kubenswrapper[4957]: I1206 06:04:21.930703 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:04:22 crc kubenswrapper[4957]: I1206 06:04:22.002745 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:04:22 crc kubenswrapper[4957]: I1206 06:04:22.127107 4957 generic.go:334] "Generic (PLEG): container finished" podID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerID="c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e" exitCode=0 Dec 06 06:04:22 crc kubenswrapper[4957]: I1206 06:04:22.127203 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerDied","Data":"c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e"} Dec 06 06:04:22 crc kubenswrapper[4957]: I1206 06:04:22.128708 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerStarted","Data":"3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713"} Dec 06 06:04:22 crc kubenswrapper[4957]: I1206 06:04:22.131778 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjk65" event={"ID":"52576ee1-8b4e-4ec9-906d-8c90e0209a7c","Type":"ContainerStarted","Data":"a04c716cf3542fb8a0b35743a42c7d764e1410a2ddc3e3fae47e8d9473b224f9"} Dec 06 06:04:22 crc kubenswrapper[4957]: I1206 06:04:22.174915 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fjk65" podStartSLOduration=9.008000978 podStartE2EDuration="17.174897742s" podCreationTimestamp="2025-12-06 06:04:05 +0000 UTC" firstStartedPulling="2025-12-06 06:04:13.018424215 +0000 UTC m=+1492.668691847" lastFinishedPulling="2025-12-06 06:04:21.185320979 +0000 UTC m=+1500.835588611" observedRunningTime="2025-12-06 06:04:22.171467918 +0000 UTC m=+1501.821735550" watchObservedRunningTime="2025-12-06 06:04:22.174897742 +0000 UTC m=+1501.825165384" Dec 06 06:04:22 crc kubenswrapper[4957]: I1206 06:04:22.717443 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdh8p"] Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.145480 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sdh8p" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="registry-server" containerID="cri-o://e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5" gracePeriod=2 Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.146046 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerStarted","Data":"4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762"} Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.529650 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.680397 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxcq6\" (UniqueName: \"kubernetes.io/projected/a497f65e-6269-468a-99e6-bad35bcb75b8-kube-api-access-hxcq6\") pod \"a497f65e-6269-468a-99e6-bad35bcb75b8\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.680735 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-utilities\") pod \"a497f65e-6269-468a-99e6-bad35bcb75b8\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.681123 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-catalog-content\") pod \"a497f65e-6269-468a-99e6-bad35bcb75b8\" (UID: \"a497f65e-6269-468a-99e6-bad35bcb75b8\") " Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.681365 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-utilities" (OuterVolumeSpecName: "utilities") pod "a497f65e-6269-468a-99e6-bad35bcb75b8" (UID: "a497f65e-6269-468a-99e6-bad35bcb75b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.683044 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.706712 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a497f65e-6269-468a-99e6-bad35bcb75b8-kube-api-access-hxcq6" (OuterVolumeSpecName: "kube-api-access-hxcq6") pod "a497f65e-6269-468a-99e6-bad35bcb75b8" (UID: "a497f65e-6269-468a-99e6-bad35bcb75b8"). InnerVolumeSpecName "kube-api-access-hxcq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.789512 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxcq6\" (UniqueName: \"kubernetes.io/projected/a497f65e-6269-468a-99e6-bad35bcb75b8-kube-api-access-hxcq6\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.795053 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a497f65e-6269-468a-99e6-bad35bcb75b8" (UID: "a497f65e-6269-468a-99e6-bad35bcb75b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:23 crc kubenswrapper[4957]: I1206 06:04:23.891464 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a497f65e-6269-468a-99e6-bad35bcb75b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.156745 4957 generic.go:334] "Generic (PLEG): container finished" podID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerID="e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5" exitCode=0 Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.156809 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh8p" event={"ID":"a497f65e-6269-468a-99e6-bad35bcb75b8","Type":"ContainerDied","Data":"e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5"} Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.156848 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh8p" event={"ID":"a497f65e-6269-468a-99e6-bad35bcb75b8","Type":"ContainerDied","Data":"f0e6bcd5dff005dd8cb38110399764c3571acfdc0ae557f41253f23754674eab"} Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.156865 4957 scope.go:117] "RemoveContainer" containerID="e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.158003 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh8p" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.159551 4957 generic.go:334] "Generic (PLEG): container finished" podID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerID="4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762" exitCode=0 Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.159604 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerDied","Data":"4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762"} Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.166369 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerStarted","Data":"a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934"} Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.198282 4957 scope.go:117] "RemoveContainer" containerID="de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.214012 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdh8p"] Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.223327 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sdh8p"] Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.224670 4957 scope.go:117] "RemoveContainer" containerID="98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.292263 4957 scope.go:117] "RemoveContainer" containerID="e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5" Dec 06 06:04:24 crc kubenswrapper[4957]: E1206 06:04:24.292844 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5\": container with ID starting with e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5 not found: ID does not exist" containerID="e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.293059 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5"} err="failed to get container status \"e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5\": rpc error: code = NotFound desc = could not find container \"e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5\": container with ID starting with e60736386d3de2eb0b72adc0cfe3408cf5427ba9d8c3c7af361630b28462cce5 not found: ID does not exist" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.293153 4957 scope.go:117] "RemoveContainer" containerID="de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea" Dec 06 06:04:24 crc kubenswrapper[4957]: E1206 06:04:24.293602 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea\": container with ID starting with de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea not found: ID does not exist" containerID="de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.293685 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea"} err="failed to get container status \"de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea\": rpc error: code = NotFound desc = could not find container \"de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea\": container with ID starting with de0633e6deff0522ea937c1cfca578e9808626a3ec78dcfd5082cea7d83c11ea not found: ID does not exist" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.293746 4957 scope.go:117] "RemoveContainer" containerID="98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86" Dec 06 06:04:24 crc kubenswrapper[4957]: E1206 06:04:24.294383 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86\": container with ID starting with 98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86 not found: ID does not exist" containerID="98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.294458 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86"} err="failed to get container status \"98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86\": rpc error: code = NotFound desc = could not find container \"98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86\": container with ID starting with 98fdef82b740a1f3db0df9a458645b552e162e0a6ac74a3741d71bd9b42e8d86 not found: ID does not exist" Dec 06 06:04:24 crc kubenswrapper[4957]: I1206 06:04:24.688020 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" path="/var/lib/kubelet/pods/a497f65e-6269-468a-99e6-bad35bcb75b8/volumes" Dec 06 06:04:25 crc kubenswrapper[4957]: I1206 06:04:25.180649 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerStarted","Data":"4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767"} Dec 06 06:04:25 crc kubenswrapper[4957]: I1206 06:04:25.210326 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4cp5j" podStartSLOduration=5.724818733 podStartE2EDuration="8.21029914s" podCreationTimestamp="2025-12-06 06:04:17 +0000 UTC" firstStartedPulling="2025-12-06 06:04:22.128791954 +0000 UTC m=+1501.779059586" lastFinishedPulling="2025-12-06 06:04:24.614272351 +0000 UTC m=+1504.264539993" observedRunningTime="2025-12-06 06:04:25.208969055 +0000 UTC m=+1504.859236707" watchObservedRunningTime="2025-12-06 06:04:25.21029914 +0000 UTC m=+1504.860566782" Dec 06 06:04:26 crc kubenswrapper[4957]: I1206 06:04:26.174965 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:26 crc kubenswrapper[4957]: I1206 06:04:26.175321 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:26 crc kubenswrapper[4957]: I1206 06:04:26.240130 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerStarted","Data":"87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c"} Dec 06 06:04:26 crc kubenswrapper[4957]: I1206 06:04:26.241560 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:04:26 crc kubenswrapper[4957]: I1206 06:04:26.277717 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:26 crc kubenswrapper[4957]: I1206 06:04:26.281802 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.776379543 podStartE2EDuration="11.281784988s" podCreationTimestamp="2025-12-06 06:04:15 +0000 UTC" firstStartedPulling="2025-12-06 06:04:16.210050148 +0000 UTC m=+1495.860317780" lastFinishedPulling="2025-12-06 06:04:25.715455583 +0000 UTC m=+1505.365723225" observedRunningTime="2025-12-06 06:04:26.279426004 +0000 UTC m=+1505.929693646" watchObservedRunningTime="2025-12-06 06:04:26.281784988 +0000 UTC m=+1505.932052630" Dec 06 06:04:26 crc kubenswrapper[4957]: I1206 06:04:26.359653 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:27 crc kubenswrapper[4957]: I1206 06:04:27.926854 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:27 crc kubenswrapper[4957]: I1206 06:04:27.928385 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:27 crc kubenswrapper[4957]: I1206 06:04:27.991981 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.132330 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fjk65"] Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.132645 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fjk65" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="registry-server" containerID="cri-o://a04c716cf3542fb8a0b35743a42c7d764e1410a2ddc3e3fae47e8d9473b224f9" gracePeriod=2 Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.272666 4957 generic.go:334] "Generic (PLEG): container finished" podID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerID="a04c716cf3542fb8a0b35743a42c7d764e1410a2ddc3e3fae47e8d9473b224f9" exitCode=0 Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.272757 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjk65" event={"ID":"52576ee1-8b4e-4ec9-906d-8c90e0209a7c","Type":"ContainerDied","Data":"a04c716cf3542fb8a0b35743a42c7d764e1410a2ddc3e3fae47e8d9473b224f9"} Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.359877 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.668097 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.804685 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-utilities\") pod \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.804819 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-catalog-content\") pod \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.804935 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mftc\" (UniqueName: \"kubernetes.io/projected/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-kube-api-access-6mftc\") pod \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\" (UID: \"52576ee1-8b4e-4ec9-906d-8c90e0209a7c\") " Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.805477 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-utilities" (OuterVolumeSpecName: "utilities") pod "52576ee1-8b4e-4ec9-906d-8c90e0209a7c" (UID: "52576ee1-8b4e-4ec9-906d-8c90e0209a7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.811123 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.814119 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-kube-api-access-6mftc" (OuterVolumeSpecName: "kube-api-access-6mftc") pod "52576ee1-8b4e-4ec9-906d-8c90e0209a7c" (UID: "52576ee1-8b4e-4ec9-906d-8c90e0209a7c"). InnerVolumeSpecName "kube-api-access-6mftc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.882294 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52576ee1-8b4e-4ec9-906d-8c90e0209a7c" (UID: "52576ee1-8b4e-4ec9-906d-8c90e0209a7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.912609 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:29 crc kubenswrapper[4957]: I1206 06:04:29.912649 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mftc\" (UniqueName: \"kubernetes.io/projected/52576ee1-8b4e-4ec9-906d-8c90e0209a7c-kube-api-access-6mftc\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.289244 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjk65" Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.289403 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjk65" event={"ID":"52576ee1-8b4e-4ec9-906d-8c90e0209a7c","Type":"ContainerDied","Data":"83215b9c5517f690fd3b16c9dd0886b64102ef7dfbff40eee6e5d9721674cee7"} Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.289455 4957 scope.go:117] "RemoveContainer" containerID="a04c716cf3542fb8a0b35743a42c7d764e1410a2ddc3e3fae47e8d9473b224f9" Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.316407 4957 scope.go:117] "RemoveContainer" containerID="a6df5585144e16ecba6e33db4c08823fad841a2564c557c5a8694006387bbbf8" Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.344388 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fjk65"] Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.350657 4957 scope.go:117] "RemoveContainer" containerID="031dc5b0afdd9c48a1f6c8fcbda785ba07bec4be442e51587758cbaa02250e2a" Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.352020 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fjk65"] Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.524417 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cp5j"] Dec 06 06:04:30 crc kubenswrapper[4957]: I1206 06:04:30.673868 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" path="/var/lib/kubelet/pods/52576ee1-8b4e-4ec9-906d-8c90e0209a7c/volumes" Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.304813 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4cp5j" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="registry-server" containerID="cri-o://4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767" gracePeriod=2 Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.767127 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.966329 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r2pf\" (UniqueName: \"kubernetes.io/projected/14414c4f-7e4f-4dfb-b44c-4a3427c36675-kube-api-access-5r2pf\") pod \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.966434 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-catalog-content\") pod \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.966553 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-utilities\") pod \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\" (UID: \"14414c4f-7e4f-4dfb-b44c-4a3427c36675\") " Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.967885 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-utilities" (OuterVolumeSpecName: "utilities") pod "14414c4f-7e4f-4dfb-b44c-4a3427c36675" (UID: "14414c4f-7e4f-4dfb-b44c-4a3427c36675"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.979125 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14414c4f-7e4f-4dfb-b44c-4a3427c36675-kube-api-access-5r2pf" (OuterVolumeSpecName: "kube-api-access-5r2pf") pod "14414c4f-7e4f-4dfb-b44c-4a3427c36675" (UID: "14414c4f-7e4f-4dfb-b44c-4a3427c36675"). InnerVolumeSpecName "kube-api-access-5r2pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:31 crc kubenswrapper[4957]: I1206 06:04:31.988134 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14414c4f-7e4f-4dfb-b44c-4a3427c36675" (UID: "14414c4f-7e4f-4dfb-b44c-4a3427c36675"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.069118 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r2pf\" (UniqueName: \"kubernetes.io/projected/14414c4f-7e4f-4dfb-b44c-4a3427c36675-kube-api-access-5r2pf\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.069161 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.069175 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14414c4f-7e4f-4dfb-b44c-4a3427c36675-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.317994 4957 generic.go:334] "Generic (PLEG): container finished" podID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerID="4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767" exitCode=0 Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.318176 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cp5j" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.318170 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerDied","Data":"4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767"} Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.318488 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cp5j" event={"ID":"14414c4f-7e4f-4dfb-b44c-4a3427c36675","Type":"ContainerDied","Data":"5aa67941d447871d63900142843c6fa7415ec2ed64d49b83b97d49ff650c5d5a"} Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.318610 4957 scope.go:117] "RemoveContainer" containerID="4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.320025 4957 generic.go:334] "Generic (PLEG): container finished" podID="e0cdc965-475d-4d47-b875-9ade77089ae6" containerID="4f2be6a8216ab58840f826fe9a624980c002648bb8b1059ae92ce567ba4c7cfd" exitCode=0 Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.320070 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" event={"ID":"e0cdc965-475d-4d47-b875-9ade77089ae6","Type":"ContainerDied","Data":"4f2be6a8216ab58840f826fe9a624980c002648bb8b1059ae92ce567ba4c7cfd"} Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.366028 4957 scope.go:117] "RemoveContainer" containerID="4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.384168 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cp5j"] Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.393686 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cp5j"] Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.403780 4957 scope.go:117] "RemoveContainer" containerID="c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.479557 4957 scope.go:117] "RemoveContainer" containerID="4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767" Dec 06 06:04:32 crc kubenswrapper[4957]: E1206 06:04:32.480210 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767\": container with ID starting with 4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767 not found: ID does not exist" containerID="4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.480258 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767"} err="failed to get container status \"4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767\": rpc error: code = NotFound desc = could not find container \"4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767\": container with ID starting with 4ddc11b8bc87701026acc0e6a723f92b7a348dd50a0e789e0ab1665b157d2767 not found: ID does not exist" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.480286 4957 scope.go:117] "RemoveContainer" containerID="4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762" Dec 06 06:04:32 crc kubenswrapper[4957]: E1206 06:04:32.480708 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762\": container with ID starting with 4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762 not found: ID does not exist" containerID="4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.480759 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762"} err="failed to get container status \"4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762\": rpc error: code = NotFound desc = could not find container \"4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762\": container with ID starting with 4e23e00b8218261d8ff08f3bb1f340e8cc33d9104fd39f3e8663189336a19762 not found: ID does not exist" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.480793 4957 scope.go:117] "RemoveContainer" containerID="c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e" Dec 06 06:04:32 crc kubenswrapper[4957]: E1206 06:04:32.481229 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e\": container with ID starting with c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e not found: ID does not exist" containerID="c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.481256 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e"} err="failed to get container status \"c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e\": rpc error: code = NotFound desc = could not find container \"c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e\": container with ID starting with c46a528b3b00b880be5a1889c514d9845888a2268d1ca0665c6d75c86d6fad8e not found: ID does not exist" Dec 06 06:04:32 crc kubenswrapper[4957]: I1206 06:04:32.674657 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" path="/var/lib/kubelet/pods/14414c4f-7e4f-4dfb-b44c-4a3427c36675/volumes" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.676372 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.804410 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-config-data\") pod \"e0cdc965-475d-4d47-b875-9ade77089ae6\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.804726 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w452r\" (UniqueName: \"kubernetes.io/projected/e0cdc965-475d-4d47-b875-9ade77089ae6-kube-api-access-w452r\") pod \"e0cdc965-475d-4d47-b875-9ade77089ae6\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.804842 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-combined-ca-bundle\") pod \"e0cdc965-475d-4d47-b875-9ade77089ae6\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.804927 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-scripts\") pod \"e0cdc965-475d-4d47-b875-9ade77089ae6\" (UID: \"e0cdc965-475d-4d47-b875-9ade77089ae6\") " Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.817235 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-scripts" (OuterVolumeSpecName: "scripts") pod "e0cdc965-475d-4d47-b875-9ade77089ae6" (UID: "e0cdc965-475d-4d47-b875-9ade77089ae6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.817403 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0cdc965-475d-4d47-b875-9ade77089ae6-kube-api-access-w452r" (OuterVolumeSpecName: "kube-api-access-w452r") pod "e0cdc965-475d-4d47-b875-9ade77089ae6" (UID: "e0cdc965-475d-4d47-b875-9ade77089ae6"). InnerVolumeSpecName "kube-api-access-w452r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.832460 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-config-data" (OuterVolumeSpecName: "config-data") pod "e0cdc965-475d-4d47-b875-9ade77089ae6" (UID: "e0cdc965-475d-4d47-b875-9ade77089ae6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.852949 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0cdc965-475d-4d47-b875-9ade77089ae6" (UID: "e0cdc965-475d-4d47-b875-9ade77089ae6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.908903 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.908950 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.908967 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w452r\" (UniqueName: \"kubernetes.io/projected/e0cdc965-475d-4d47-b875-9ade77089ae6-kube-api-access-w452r\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:33 crc kubenswrapper[4957]: I1206 06:04:33.908980 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0cdc965-475d-4d47-b875-9ade77089ae6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.347950 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" event={"ID":"e0cdc965-475d-4d47-b875-9ade77089ae6","Type":"ContainerDied","Data":"0f2c4e4d68692e2ef70694bae199672a1d71514ee398d7994624349a79864c99"} Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.348019 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f2c4e4d68692e2ef70694bae199672a1d71514ee398d7994624349a79864c99" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.348121 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ccb8p" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.486568 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487285 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="extract-utilities" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487326 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="extract-utilities" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487362 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="extract-utilities" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487375 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="extract-utilities" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487441 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487455 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487479 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487492 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487512 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cdc965-475d-4d47-b875-9ade77089ae6" containerName="nova-cell0-conductor-db-sync" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487524 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cdc965-475d-4d47-b875-9ade77089ae6" containerName="nova-cell0-conductor-db-sync" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487546 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="extract-content" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487559 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="extract-content" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487578 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487590 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487626 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="extract-utilities" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487637 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="extract-utilities" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487667 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="extract-content" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487680 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="extract-content" Dec 06 06:04:34 crc kubenswrapper[4957]: E1206 06:04:34.487709 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="extract-content" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.487720 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="extract-content" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.488174 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="14414c4f-7e4f-4dfb-b44c-4a3427c36675" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.488203 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cdc965-475d-4d47-b875-9ade77089ae6" containerName="nova-cell0-conductor-db-sync" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.488237 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a497f65e-6269-468a-99e6-bad35bcb75b8" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.488281 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="52576ee1-8b4e-4ec9-906d-8c90e0209a7c" containerName="registry-server" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.489560 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.494009 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wz64c" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.499731 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.505671 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.624745 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ce84760-5e40-4ea1-8bb2-81dc593c757a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.625060 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhjnp\" (UniqueName: \"kubernetes.io/projected/9ce84760-5e40-4ea1-8bb2-81dc593c757a-kube-api-access-dhjnp\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.625329 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ce84760-5e40-4ea1-8bb2-81dc593c757a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.728289 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ce84760-5e40-4ea1-8bb2-81dc593c757a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.728803 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ce84760-5e40-4ea1-8bb2-81dc593c757a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.729105 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhjnp\" (UniqueName: \"kubernetes.io/projected/9ce84760-5e40-4ea1-8bb2-81dc593c757a-kube-api-access-dhjnp\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.734156 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ce84760-5e40-4ea1-8bb2-81dc593c757a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.736804 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ce84760-5e40-4ea1-8bb2-81dc593c757a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.754605 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhjnp\" (UniqueName: \"kubernetes.io/projected/9ce84760-5e40-4ea1-8bb2-81dc593c757a-kube-api-access-dhjnp\") pod \"nova-cell0-conductor-0\" (UID: \"9ce84760-5e40-4ea1-8bb2-81dc593c757a\") " pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:34 crc kubenswrapper[4957]: I1206 06:04:34.813806 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:35 crc kubenswrapper[4957]: I1206 06:04:35.403941 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 06:04:36 crc kubenswrapper[4957]: I1206 06:04:36.383718 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9ce84760-5e40-4ea1-8bb2-81dc593c757a","Type":"ContainerStarted","Data":"14e23120a4db8c98048516f70ac1649806421299af3a3cd8544665e57ca83887"} Dec 06 06:04:36 crc kubenswrapper[4957]: I1206 06:04:36.384085 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9ce84760-5e40-4ea1-8bb2-81dc593c757a","Type":"ContainerStarted","Data":"8b73534f75090ba5b87d2533af3642c1b05e418b278f383a299aa2d6962aaa5e"} Dec 06 06:04:36 crc kubenswrapper[4957]: I1206 06:04:36.384101 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:36 crc kubenswrapper[4957]: I1206 06:04:36.405828 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.4058092 podStartE2EDuration="2.4058092s" podCreationTimestamp="2025-12-06 06:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:04:36.397238559 +0000 UTC m=+1516.047506201" watchObservedRunningTime="2025-12-06 06:04:36.4058092 +0000 UTC m=+1516.056076832" Dec 06 06:04:40 crc kubenswrapper[4957]: I1206 06:04:40.180749 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:04:40 crc kubenswrapper[4957]: I1206 06:04:40.181380 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:04:44 crc kubenswrapper[4957]: I1206 06:04:44.861535 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.330660 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-hplhf"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.345146 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.368518 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.368632 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.394087 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hplhf"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.455374 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-config-data\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.455459 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64nj5\" (UniqueName: \"kubernetes.io/projected/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-kube-api-access-64nj5\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.455496 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.455526 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-scripts\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.553888 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.555485 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.557278 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64nj5\" (UniqueName: \"kubernetes.io/projected/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-kube-api-access-64nj5\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.557324 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.557357 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-scripts\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.557432 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-config-data\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.563682 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.564196 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.564380 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-scripts\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.572549 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-config-data\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.598442 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.622589 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64nj5\" (UniqueName: \"kubernetes.io/projected/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-kube-api-access-64nj5\") pod \"nova-cell0-cell-mapping-hplhf\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.658937 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.658986 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0501ce8c-285f-4d42-abaa-1c7252324b0e-logs\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.659034 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pb5f\" (UniqueName: \"kubernetes.io/projected/0501ce8c-285f-4d42-abaa-1c7252324b0e-kube-api-access-4pb5f\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.659064 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-config-data\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.680264 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.681597 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.685617 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.697262 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.706665 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.756147 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.760349 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pb5f\" (UniqueName: \"kubernetes.io/projected/0501ce8c-285f-4d42-abaa-1c7252324b0e-kube-api-access-4pb5f\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.760433 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-config-data\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.760643 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.760668 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0501ce8c-285f-4d42-abaa-1c7252324b0e-logs\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.761137 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0501ce8c-285f-4d42-abaa-1c7252324b0e-logs\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.783565 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.803478 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.807979 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-config-data\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.813638 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.824581 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.847101 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.851475 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pb5f\" (UniqueName: \"kubernetes.io/projected/0501ce8c-285f-4d42-abaa-1c7252324b0e-kube-api-access-4pb5f\") pod \"nova-api-0\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " pod="openstack/nova-api-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.879985 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.880068 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh9cb\" (UniqueName: \"kubernetes.io/projected/82cd2252-346e-4456-b059-193abd84d10d-kube-api-access-zh9cb\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.880199 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.961382 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.966385 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.972775 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.984825 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.984897 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.984940 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft2l7\" (UniqueName: \"kubernetes.io/projected/61d6eca8-07f2-473b-9e1e-34756cd9e33a-kube-api-access-ft2l7\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.985019 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh9cb\" (UniqueName: \"kubernetes.io/projected/82cd2252-346e-4456-b059-193abd84d10d-kube-api-access-zh9cb\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.985057 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-config-data\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:45 crc kubenswrapper[4957]: I1206 06:04:45.985184 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.002527 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.009382 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.011990 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.021368 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-vx2p7"] Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.022559 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.022757 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.043946 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh9cb\" (UniqueName: \"kubernetes.io/projected/82cd2252-346e-4456-b059-193abd84d10d-kube-api-access-zh9cb\") pod \"nova-cell1-novncproxy-0\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.077825 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-vx2p7"] Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.087187 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-config-data\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.087297 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.087319 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz4fd\" (UniqueName: \"kubernetes.io/projected/5c757adb-1ed5-40b7-a923-79b89d269f1e-kube-api-access-lz4fd\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.087348 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c757adb-1ed5-40b7-a923-79b89d269f1e-logs\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.087394 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-config-data\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.087415 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.087445 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft2l7\" (UniqueName: \"kubernetes.io/projected/61d6eca8-07f2-473b-9e1e-34756cd9e33a-kube-api-access-ft2l7\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.093918 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.094940 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-config-data\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.103674 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft2l7\" (UniqueName: \"kubernetes.io/projected/61d6eca8-07f2-473b-9e1e-34756cd9e33a-kube-api-access-ft2l7\") pod \"nova-scheduler-0\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " pod="openstack/nova-scheduler-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.177962 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189284 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-config\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189360 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-config-data\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189391 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd2nk\" (UniqueName: \"kubernetes.io/projected/296035c8-c16e-4a85-b712-78f8ee8f70cd-kube-api-access-pd2nk\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189435 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189475 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189499 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189578 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-svc\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189631 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189655 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz4fd\" (UniqueName: \"kubernetes.io/projected/5c757adb-1ed5-40b7-a923-79b89d269f1e-kube-api-access-lz4fd\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.189695 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c757adb-1ed5-40b7-a923-79b89d269f1e-logs\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.190137 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c757adb-1ed5-40b7-a923-79b89d269f1e-logs\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.195798 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.202313 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-config-data\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.203362 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.209284 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz4fd\" (UniqueName: \"kubernetes.io/projected/5c757adb-1ed5-40b7-a923-79b89d269f1e-kube-api-access-lz4fd\") pod \"nova-metadata-0\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.293813 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd2nk\" (UniqueName: \"kubernetes.io/projected/296035c8-c16e-4a85-b712-78f8ee8f70cd-kube-api-access-pd2nk\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.293893 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.293927 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.293942 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.294007 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-svc\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.294083 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-config\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.294995 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.295201 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-config\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.295748 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.297225 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.299636 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.301421 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-svc\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.321753 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd2nk\" (UniqueName: \"kubernetes.io/projected/296035c8-c16e-4a85-b712-78f8ee8f70cd-kube-api-access-pd2nk\") pod \"dnsmasq-dns-757b4f8459-vx2p7\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.369267 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.464572 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hplhf"] Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.623454 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.895899 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xjg58"] Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.907569 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.916622 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 06 06:04:46 crc kubenswrapper[4957]: I1206 06:04:46.917044 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 06:04:46 crc kubenswrapper[4957]: W1206 06:04:46.994215 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61d6eca8_07f2_473b_9e1e_34756cd9e33a.slice/crio-481ff8e4f9f210afe0d83cda1282458e217db4e6d0223be09be3603dc2f9ff64 WatchSource:0}: Error finding container 481ff8e4f9f210afe0d83cda1282458e217db4e6d0223be09be3603dc2f9ff64: Status 404 returned error can't find the container with id 481ff8e4f9f210afe0d83cda1282458e217db4e6d0223be09be3603dc2f9ff64 Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.018243 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xjg58"] Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.026161 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-config-data\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.026230 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57fq8\" (UniqueName: \"kubernetes.io/projected/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-kube-api-access-57fq8\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.026288 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.026370 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-scripts\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.066195 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.083203 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.127530 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-scripts\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.127629 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-config-data\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.127663 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57fq8\" (UniqueName: \"kubernetes.io/projected/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-kube-api-access-57fq8\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.127702 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.135255 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-scripts\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.136023 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.141688 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-config-data\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.165297 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57fq8\" (UniqueName: \"kubernetes.io/projected/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-kube-api-access-57fq8\") pod \"nova-cell1-conductor-db-sync-xjg58\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.227263 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:04:47 crc kubenswrapper[4957]: W1206 06:04:47.245648 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod296035c8_c16e_4a85_b712_78f8ee8f70cd.slice/crio-90e931ae681b26eeaf323294c2cd78dbb9c89c69bfff55f1842d9595350f33dc WatchSource:0}: Error finding container 90e931ae681b26eeaf323294c2cd78dbb9c89c69bfff55f1842d9595350f33dc: Status 404 returned error can't find the container with id 90e931ae681b26eeaf323294c2cd78dbb9c89c69bfff55f1842d9595350f33dc Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.254211 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.256154 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-vx2p7"] Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.616190 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82cd2252-346e-4456-b059-193abd84d10d","Type":"ContainerStarted","Data":"68c25519ceef1a60ed88408e219dcdd1031410d246ea0547a3a96602f80759d8"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.618326 4957 generic.go:334] "Generic (PLEG): container finished" podID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerID="7c19bfc973175473ca1d34592555409e2a43abeda269082efb9ccbeed0ba1223" exitCode=0 Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.618420 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" event={"ID":"296035c8-c16e-4a85-b712-78f8ee8f70cd","Type":"ContainerDied","Data":"7c19bfc973175473ca1d34592555409e2a43abeda269082efb9ccbeed0ba1223"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.618443 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" event={"ID":"296035c8-c16e-4a85-b712-78f8ee8f70cd","Type":"ContainerStarted","Data":"90e931ae681b26eeaf323294c2cd78dbb9c89c69bfff55f1842d9595350f33dc"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.622286 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0501ce8c-285f-4d42-abaa-1c7252324b0e","Type":"ContainerStarted","Data":"a4627e08cbbd1768642639481054d496676167d66f4573b46e1b6fbf0e7547ec"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.629215 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hplhf" event={"ID":"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8","Type":"ContainerStarted","Data":"d0a9a72d7eef5d65a50afd7b21162ef1e895c03133cf87e5d26afa27076760a4"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.629266 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hplhf" event={"ID":"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8","Type":"ContainerStarted","Data":"33d330c419a5e81f1e8a1a3b487e8d4980795bf5e26d8590a23705505fb1c589"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.632512 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61d6eca8-07f2-473b-9e1e-34756cd9e33a","Type":"ContainerStarted","Data":"481ff8e4f9f210afe0d83cda1282458e217db4e6d0223be09be3603dc2f9ff64"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.634100 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c757adb-1ed5-40b7-a923-79b89d269f1e","Type":"ContainerStarted","Data":"7970c0ae2e19fe0826a397c6f2c4b14d52a964340e63a5ea8cf24d358178994c"} Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.725383 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-hplhf" podStartSLOduration=2.725363464 podStartE2EDuration="2.725363464s" podCreationTimestamp="2025-12-06 06:04:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:04:47.66529834 +0000 UTC m=+1527.315565972" watchObservedRunningTime="2025-12-06 06:04:47.725363464 +0000 UTC m=+1527.375631086" Dec 06 06:04:47 crc kubenswrapper[4957]: I1206 06:04:47.733948 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xjg58"] Dec 06 06:04:47 crc kubenswrapper[4957]: W1206 06:04:47.751774 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6db5d8d7_7fc2_437c_a7bb_5c3630eca4f3.slice/crio-d0b2c0b3692a16934278aa8feaea8497c9042680d05b2fc97b21325ef55092f6 WatchSource:0}: Error finding container d0b2c0b3692a16934278aa8feaea8497c9042680d05b2fc97b21325ef55092f6: Status 404 returned error can't find the container with id d0b2c0b3692a16934278aa8feaea8497c9042680d05b2fc97b21325ef55092f6 Dec 06 06:04:48 crc kubenswrapper[4957]: I1206 06:04:48.658255 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xjg58" event={"ID":"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3","Type":"ContainerStarted","Data":"ae82e43bea073eaff0c99228d7792633beab7e0865f7a8b12ba11fd1cd1f1c62"} Dec 06 06:04:48 crc kubenswrapper[4957]: I1206 06:04:48.658605 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xjg58" event={"ID":"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3","Type":"ContainerStarted","Data":"d0b2c0b3692a16934278aa8feaea8497c9042680d05b2fc97b21325ef55092f6"} Dec 06 06:04:48 crc kubenswrapper[4957]: I1206 06:04:48.665520 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" event={"ID":"296035c8-c16e-4a85-b712-78f8ee8f70cd","Type":"ContainerStarted","Data":"8867c86d2b2990d2807f29679c6133960ebd025317ef5b7906227a45aed90f06"} Dec 06 06:04:48 crc kubenswrapper[4957]: I1206 06:04:48.698487 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xjg58" podStartSLOduration=2.698454391 podStartE2EDuration="2.698454391s" podCreationTimestamp="2025-12-06 06:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:04:48.684919295 +0000 UTC m=+1528.335186927" watchObservedRunningTime="2025-12-06 06:04:48.698454391 +0000 UTC m=+1528.348722023" Dec 06 06:04:48 crc kubenswrapper[4957]: I1206 06:04:48.703245 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:48 crc kubenswrapper[4957]: I1206 06:04:48.713735 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" podStartSLOduration=3.713714493 podStartE2EDuration="3.713714493s" podCreationTimestamp="2025-12-06 06:04:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:04:48.711446182 +0000 UTC m=+1528.361713834" watchObservedRunningTime="2025-12-06 06:04:48.713714493 +0000 UTC m=+1528.363982125" Dec 06 06:04:49 crc kubenswrapper[4957]: I1206 06:04:49.163888 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:04:49 crc kubenswrapper[4957]: I1206 06:04:49.223114 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.726067 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c757adb-1ed5-40b7-a923-79b89d269f1e","Type":"ContainerStarted","Data":"4c8b6674d15d7b91045bb18c121ac04a1270deda62a69aab95c9d178013e71b0"} Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.726445 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c757adb-1ed5-40b7-a923-79b89d269f1e","Type":"ContainerStarted","Data":"7833c38e87729d4fea3cf2bfc43624395dcf36d7ef229bb2c3ff57c271cac9c4"} Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.726591 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-log" containerID="cri-o://7833c38e87729d4fea3cf2bfc43624395dcf36d7ef229bb2c3ff57c271cac9c4" gracePeriod=30 Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.726718 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-metadata" containerID="cri-o://4c8b6674d15d7b91045bb18c121ac04a1270deda62a69aab95c9d178013e71b0" gracePeriod=30 Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.738434 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="82cd2252-346e-4456-b059-193abd84d10d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c6cb89a94d7d6e6ce03f464bc9ac4b7a8ed6e181d3c005801e90785499667870" gracePeriod=30 Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.738810 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82cd2252-346e-4456-b059-193abd84d10d","Type":"ContainerStarted","Data":"c6cb89a94d7d6e6ce03f464bc9ac4b7a8ed6e181d3c005801e90785499667870"} Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.745292 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0501ce8c-285f-4d42-abaa-1c7252324b0e","Type":"ContainerStarted","Data":"17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b"} Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.745359 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0501ce8c-285f-4d42-abaa-1c7252324b0e","Type":"ContainerStarted","Data":"2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e"} Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.748080 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61d6eca8-07f2-473b-9e1e-34756cd9e33a","Type":"ContainerStarted","Data":"37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e"} Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.796391 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2695152370000002 podStartE2EDuration="5.796372237s" podCreationTimestamp="2025-12-06 06:04:45 +0000 UTC" firstStartedPulling="2025-12-06 06:04:46.656902959 +0000 UTC m=+1526.307170591" lastFinishedPulling="2025-12-06 06:04:50.183759959 +0000 UTC m=+1529.834027591" observedRunningTime="2025-12-06 06:04:50.775753579 +0000 UTC m=+1530.426021221" watchObservedRunningTime="2025-12-06 06:04:50.796372237 +0000 UTC m=+1530.446639869" Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.801696 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.843922529 podStartE2EDuration="5.801674159s" podCreationTimestamp="2025-12-06 06:04:45 +0000 UTC" firstStartedPulling="2025-12-06 06:04:47.225018143 +0000 UTC m=+1526.875285775" lastFinishedPulling="2025-12-06 06:04:50.182769773 +0000 UTC m=+1529.833037405" observedRunningTime="2025-12-06 06:04:50.796365897 +0000 UTC m=+1530.446633539" watchObservedRunningTime="2025-12-06 06:04:50.801674159 +0000 UTC m=+1530.451941791" Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.819870 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.637961742 podStartE2EDuration="5.819822861s" podCreationTimestamp="2025-12-06 06:04:45 +0000 UTC" firstStartedPulling="2025-12-06 06:04:47.003195275 +0000 UTC m=+1526.653462907" lastFinishedPulling="2025-12-06 06:04:50.185056394 +0000 UTC m=+1529.835324026" observedRunningTime="2025-12-06 06:04:50.813639864 +0000 UTC m=+1530.463907506" watchObservedRunningTime="2025-12-06 06:04:50.819822861 +0000 UTC m=+1530.470090493" Dec 06 06:04:50 crc kubenswrapper[4957]: I1206 06:04:50.838468 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.726531986 podStartE2EDuration="5.838448435s" podCreationTimestamp="2025-12-06 06:04:45 +0000 UTC" firstStartedPulling="2025-12-06 06:04:47.07366729 +0000 UTC m=+1526.723934922" lastFinishedPulling="2025-12-06 06:04:50.185583739 +0000 UTC m=+1529.835851371" observedRunningTime="2025-12-06 06:04:50.836311517 +0000 UTC m=+1530.486579159" watchObservedRunningTime="2025-12-06 06:04:50.838448435 +0000 UTC m=+1530.488716067" Dec 06 06:04:51 crc kubenswrapper[4957]: I1206 06:04:51.178678 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:04:51 crc kubenswrapper[4957]: I1206 06:04:51.204137 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 06:04:51 crc kubenswrapper[4957]: I1206 06:04:51.298577 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:04:51 crc kubenswrapper[4957]: I1206 06:04:51.298627 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:04:51 crc kubenswrapper[4957]: I1206 06:04:51.760461 4957 generic.go:334] "Generic (PLEG): container finished" podID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerID="7833c38e87729d4fea3cf2bfc43624395dcf36d7ef229bb2c3ff57c271cac9c4" exitCode=143 Dec 06 06:04:51 crc kubenswrapper[4957]: I1206 06:04:51.761036 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c757adb-1ed5-40b7-a923-79b89d269f1e","Type":"ContainerDied","Data":"7833c38e87729d4fea3cf2bfc43624395dcf36d7ef229bb2c3ff57c271cac9c4"} Dec 06 06:04:52 crc kubenswrapper[4957]: I1206 06:04:52.470575 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:04:52 crc kubenswrapper[4957]: I1206 06:04:52.471087 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="33344326-7541-47d7-8b16-0df124dce272" containerName="kube-state-metrics" containerID="cri-o://2b5e3d35938d1d4b9b38309b7fec8bd84334a0f650e7a705fe5fc92b13443c2a" gracePeriod=30 Dec 06 06:04:52 crc kubenswrapper[4957]: I1206 06:04:52.779416 4957 generic.go:334] "Generic (PLEG): container finished" podID="33344326-7541-47d7-8b16-0df124dce272" containerID="2b5e3d35938d1d4b9b38309b7fec8bd84334a0f650e7a705fe5fc92b13443c2a" exitCode=2 Dec 06 06:04:52 crc kubenswrapper[4957]: I1206 06:04:52.779528 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"33344326-7541-47d7-8b16-0df124dce272","Type":"ContainerDied","Data":"2b5e3d35938d1d4b9b38309b7fec8bd84334a0f650e7a705fe5fc92b13443c2a"} Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.003561 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.064117 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg7j5\" (UniqueName: \"kubernetes.io/projected/33344326-7541-47d7-8b16-0df124dce272-kube-api-access-tg7j5\") pod \"33344326-7541-47d7-8b16-0df124dce272\" (UID: \"33344326-7541-47d7-8b16-0df124dce272\") " Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.070909 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33344326-7541-47d7-8b16-0df124dce272-kube-api-access-tg7j5" (OuterVolumeSpecName: "kube-api-access-tg7j5") pod "33344326-7541-47d7-8b16-0df124dce272" (UID: "33344326-7541-47d7-8b16-0df124dce272"). InnerVolumeSpecName "kube-api-access-tg7j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.166946 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg7j5\" (UniqueName: \"kubernetes.io/projected/33344326-7541-47d7-8b16-0df124dce272-kube-api-access-tg7j5\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.789291 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"33344326-7541-47d7-8b16-0df124dce272","Type":"ContainerDied","Data":"418f2509b76d912259234fad2184fcece35ecaeb4cf66e9bb338921883259f04"} Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.789358 4957 scope.go:117] "RemoveContainer" containerID="2b5e3d35938d1d4b9b38309b7fec8bd84334a0f650e7a705fe5fc92b13443c2a" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.789361 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.831556 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.845408 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.859863 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:04:53 crc kubenswrapper[4957]: E1206 06:04:53.860537 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33344326-7541-47d7-8b16-0df124dce272" containerName="kube-state-metrics" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.860572 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="33344326-7541-47d7-8b16-0df124dce272" containerName="kube-state-metrics" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.860960 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="33344326-7541-47d7-8b16-0df124dce272" containerName="kube-state-metrics" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.862054 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.870306 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.872314 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 06 06:04:53 crc kubenswrapper[4957]: I1206 06:04:53.877726 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.001731 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.002112 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.002193 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.002232 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxltm\" (UniqueName: \"kubernetes.io/projected/6a476151-cb3f-4502-8b9a-be049c359804-kube-api-access-dxltm\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.104352 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.104779 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxltm\" (UniqueName: \"kubernetes.io/projected/6a476151-cb3f-4502-8b9a-be049c359804-kube-api-access-dxltm\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.105420 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.106211 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.114439 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.118904 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.122720 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxltm\" (UniqueName: \"kubernetes.io/projected/6a476151-cb3f-4502-8b9a-be049c359804-kube-api-access-dxltm\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.123612 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6a476151-cb3f-4502-8b9a-be049c359804-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6a476151-cb3f-4502-8b9a-be049c359804\") " pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.188192 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.205033 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.205353 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-central-agent" containerID="cri-o://fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a" gracePeriod=30 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.209264 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="proxy-httpd" containerID="cri-o://87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c" gracePeriod=30 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.209495 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-notification-agent" containerID="cri-o://3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713" gracePeriod=30 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.215113 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="sg-core" containerID="cri-o://a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934" gracePeriod=30 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.673062 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33344326-7541-47d7-8b16-0df124dce272" path="/var/lib/kubelet/pods/33344326-7541-47d7-8b16-0df124dce272/volumes" Dec 06 06:04:54 crc kubenswrapper[4957]: W1206 06:04:54.773696 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a476151_cb3f_4502_8b9a_be049c359804.slice/crio-883849665b19bbbb84a5582901efb12854fca698e824cfbde2d4e25efecb83e8 WatchSource:0}: Error finding container 883849665b19bbbb84a5582901efb12854fca698e824cfbde2d4e25efecb83e8: Status 404 returned error can't find the container with id 883849665b19bbbb84a5582901efb12854fca698e824cfbde2d4e25efecb83e8 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.783776 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.800373 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6a476151-cb3f-4502-8b9a-be049c359804","Type":"ContainerStarted","Data":"883849665b19bbbb84a5582901efb12854fca698e824cfbde2d4e25efecb83e8"} Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.806010 4957 generic.go:334] "Generic (PLEG): container finished" podID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerID="87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c" exitCode=0 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.806038 4957 generic.go:334] "Generic (PLEG): container finished" podID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerID="a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934" exitCode=2 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.806050 4957 generic.go:334] "Generic (PLEG): container finished" podID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerID="fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a" exitCode=0 Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.806070 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerDied","Data":"87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c"} Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.806094 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerDied","Data":"a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934"} Dec 06 06:04:54 crc kubenswrapper[4957]: I1206 06:04:54.806104 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerDied","Data":"fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a"} Dec 06 06:04:55 crc kubenswrapper[4957]: I1206 06:04:55.823599 4957 generic.go:334] "Generic (PLEG): container finished" podID="d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" containerID="d0a9a72d7eef5d65a50afd7b21162ef1e895c03133cf87e5d26afa27076760a4" exitCode=0 Dec 06 06:04:55 crc kubenswrapper[4957]: I1206 06:04:55.823948 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hplhf" event={"ID":"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8","Type":"ContainerDied","Data":"d0a9a72d7eef5d65a50afd7b21162ef1e895c03133cf87e5d26afa27076760a4"} Dec 06 06:04:55 crc kubenswrapper[4957]: I1206 06:04:55.829103 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6a476151-cb3f-4502-8b9a-be049c359804","Type":"ContainerStarted","Data":"5b11b9c1b678e56ef6652dc84f34f12e449b1b6e151a3dd29d296ddc7c5cdcf8"} Dec 06 06:04:55 crc kubenswrapper[4957]: I1206 06:04:55.829877 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 06:04:55 crc kubenswrapper[4957]: I1206 06:04:55.862035 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.498651894 podStartE2EDuration="2.862014671s" podCreationTimestamp="2025-12-06 06:04:53 +0000 UTC" firstStartedPulling="2025-12-06 06:04:54.776423632 +0000 UTC m=+1534.426691274" lastFinishedPulling="2025-12-06 06:04:55.139786399 +0000 UTC m=+1534.790054051" observedRunningTime="2025-12-06 06:04:55.858373053 +0000 UTC m=+1535.508640685" watchObservedRunningTime="2025-12-06 06:04:55.862014671 +0000 UTC m=+1535.512282303" Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.024650 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.024711 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.204873 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.245000 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.373819 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.425774 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9b95f"] Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.426014 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" podUID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerName="dnsmasq-dns" containerID="cri-o://8a02407f8552fa56458c1278085e94b00bc965d3cf46bbdb8cc4c13cde53b91f" gracePeriod=10 Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.838108 4957 generic.go:334] "Generic (PLEG): container finished" podID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerID="8a02407f8552fa56458c1278085e94b00bc965d3cf46bbdb8cc4c13cde53b91f" exitCode=0 Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.838216 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" event={"ID":"f840f7f6-e5f8-4307-910f-5c30200ae1fb","Type":"ContainerDied","Data":"8a02407f8552fa56458c1278085e94b00bc965d3cf46bbdb8cc4c13cde53b91f"} Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.840521 4957 generic.go:334] "Generic (PLEG): container finished" podID="6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" containerID="ae82e43bea073eaff0c99228d7792633beab7e0865f7a8b12ba11fd1cd1f1c62" exitCode=0 Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.840722 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xjg58" event={"ID":"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3","Type":"ContainerDied","Data":"ae82e43bea073eaff0c99228d7792633beab7e0865f7a8b12ba11fd1cd1f1c62"} Dec 06 06:04:56 crc kubenswrapper[4957]: I1206 06:04:56.880551 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.108022 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.108147 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.328205 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.488551 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-combined-ca-bundle\") pod \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.488615 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64nj5\" (UniqueName: \"kubernetes.io/projected/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-kube-api-access-64nj5\") pod \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.488809 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-scripts\") pod \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.488871 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-config-data\") pod \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\" (UID: \"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.495732 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-scripts" (OuterVolumeSpecName: "scripts") pod "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" (UID: "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.496590 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.498015 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-kube-api-access-64nj5" (OuterVolumeSpecName: "kube-api-access-64nj5") pod "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" (UID: "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8"). InnerVolumeSpecName "kube-api-access-64nj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.551977 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" (UID: "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.563036 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-config-data" (OuterVolumeSpecName: "config-data") pod "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" (UID: "d31daa9e-4a19-46c6-bdd0-eff90eb42ba8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.590898 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-nb\") pod \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.590936 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-swift-storage-0\") pod \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.590962 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-config\") pod \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.591079 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch2sb\" (UniqueName: \"kubernetes.io/projected/f840f7f6-e5f8-4307-910f-5c30200ae1fb-kube-api-access-ch2sb\") pod \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.591130 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-sb\") pod \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.591179 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-svc\") pod \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\" (UID: \"f840f7f6-e5f8-4307-910f-5c30200ae1fb\") " Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.591534 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.591554 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64nj5\" (UniqueName: \"kubernetes.io/projected/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-kube-api-access-64nj5\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.591564 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.591572 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.594890 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f840f7f6-e5f8-4307-910f-5c30200ae1fb-kube-api-access-ch2sb" (OuterVolumeSpecName: "kube-api-access-ch2sb") pod "f840f7f6-e5f8-4307-910f-5c30200ae1fb" (UID: "f840f7f6-e5f8-4307-910f-5c30200ae1fb"). InnerVolumeSpecName "kube-api-access-ch2sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.654458 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f840f7f6-e5f8-4307-910f-5c30200ae1fb" (UID: "f840f7f6-e5f8-4307-910f-5c30200ae1fb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.663276 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f840f7f6-e5f8-4307-910f-5c30200ae1fb" (UID: "f840f7f6-e5f8-4307-910f-5c30200ae1fb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.666539 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f840f7f6-e5f8-4307-910f-5c30200ae1fb" (UID: "f840f7f6-e5f8-4307-910f-5c30200ae1fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.669245 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-config" (OuterVolumeSpecName: "config") pod "f840f7f6-e5f8-4307-910f-5c30200ae1fb" (UID: "f840f7f6-e5f8-4307-910f-5c30200ae1fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.671054 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f840f7f6-e5f8-4307-910f-5c30200ae1fb" (UID: "f840f7f6-e5f8-4307-910f-5c30200ae1fb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.693594 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.693641 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.693658 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.693674 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch2sb\" (UniqueName: \"kubernetes.io/projected/f840f7f6-e5f8-4307-910f-5c30200ae1fb-kube-api-access-ch2sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.693689 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.693702 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f840f7f6-e5f8-4307-910f-5c30200ae1fb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.850261 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hplhf" event={"ID":"d31daa9e-4a19-46c6-bdd0-eff90eb42ba8","Type":"ContainerDied","Data":"33d330c419a5e81f1e8a1a3b487e8d4980795bf5e26d8590a23705505fb1c589"} Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.850304 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33d330c419a5e81f1e8a1a3b487e8d4980795bf5e26d8590a23705505fb1c589" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.850364 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hplhf" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.862715 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" event={"ID":"f840f7f6-e5f8-4307-910f-5c30200ae1fb","Type":"ContainerDied","Data":"9b55a6be087026ac4dfb11a3ba631685c822870b9141e4dbce9e5472d5dfc944"} Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.862783 4957 scope.go:117] "RemoveContainer" containerID="8a02407f8552fa56458c1278085e94b00bc965d3cf46bbdb8cc4c13cde53b91f" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.863348 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9b95f" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.916427 4957 scope.go:117] "RemoveContainer" containerID="9490628da4465835c0ddccb6293ac2d15881077d9cbc74f23b692569da9a0f93" Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.939969 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9b95f"] Dec 06 06:04:57 crc kubenswrapper[4957]: I1206 06:04:57.953515 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9b95f"] Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.066728 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.067124 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-log" containerID="cri-o://2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e" gracePeriod=30 Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.067388 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-api" containerID="cri-o://17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b" gracePeriod=30 Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.089398 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.261348 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.411719 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-combined-ca-bundle\") pod \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.411950 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57fq8\" (UniqueName: \"kubernetes.io/projected/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-kube-api-access-57fq8\") pod \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.412025 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-config-data\") pod \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.412080 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-scripts\") pod \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\" (UID: \"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.417678 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-scripts" (OuterVolumeSpecName: "scripts") pod "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" (UID: "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.417953 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-kube-api-access-57fq8" (OuterVolumeSpecName: "kube-api-access-57fq8") pod "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" (UID: "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3"). InnerVolumeSpecName "kube-api-access-57fq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.451114 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" (UID: "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.462922 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-config-data" (OuterVolumeSpecName: "config-data") pod "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" (UID: "6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.514246 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.514289 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57fq8\" (UniqueName: \"kubernetes.io/projected/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-kube-api-access-57fq8\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.514307 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.514322 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.669482 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.674306 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" path="/var/lib/kubelet/pods/f840f7f6-e5f8-4307-910f-5c30200ae1fb/volumes" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821331 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-log-httpd\") pod \"97341258-bed0-4dda-9b61-bf6e9c93d375\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821660 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-combined-ca-bundle\") pod \"97341258-bed0-4dda-9b61-bf6e9c93d375\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821681 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-scripts\") pod \"97341258-bed0-4dda-9b61-bf6e9c93d375\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821721 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-sg-core-conf-yaml\") pod \"97341258-bed0-4dda-9b61-bf6e9c93d375\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821724 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "97341258-bed0-4dda-9b61-bf6e9c93d375" (UID: "97341258-bed0-4dda-9b61-bf6e9c93d375"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821749 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-config-data\") pod \"97341258-bed0-4dda-9b61-bf6e9c93d375\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821780 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-run-httpd\") pod \"97341258-bed0-4dda-9b61-bf6e9c93d375\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.821879 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44fz6\" (UniqueName: \"kubernetes.io/projected/97341258-bed0-4dda-9b61-bf6e9c93d375-kube-api-access-44fz6\") pod \"97341258-bed0-4dda-9b61-bf6e9c93d375\" (UID: \"97341258-bed0-4dda-9b61-bf6e9c93d375\") " Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.822028 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "97341258-bed0-4dda-9b61-bf6e9c93d375" (UID: "97341258-bed0-4dda-9b61-bf6e9c93d375"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.822327 4957 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.822342 4957 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/97341258-bed0-4dda-9b61-bf6e9c93d375-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.827914 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-scripts" (OuterVolumeSpecName: "scripts") pod "97341258-bed0-4dda-9b61-bf6e9c93d375" (UID: "97341258-bed0-4dda-9b61-bf6e9c93d375"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.828377 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97341258-bed0-4dda-9b61-bf6e9c93d375-kube-api-access-44fz6" (OuterVolumeSpecName: "kube-api-access-44fz6") pod "97341258-bed0-4dda-9b61-bf6e9c93d375" (UID: "97341258-bed0-4dda-9b61-bf6e9c93d375"). InnerVolumeSpecName "kube-api-access-44fz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.886565 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "97341258-bed0-4dda-9b61-bf6e9c93d375" (UID: "97341258-bed0-4dda-9b61-bf6e9c93d375"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.914297 4957 generic.go:334] "Generic (PLEG): container finished" podID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerID="3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713" exitCode=0 Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.914437 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.919209 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerDied","Data":"3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713"} Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.919300 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"97341258-bed0-4dda-9b61-bf6e9c93d375","Type":"ContainerDied","Data":"cbb0c6fa736e2dba51275abf9b3fd30b0e4b428807c8bb5a4c12f3ebebac7b34"} Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.919318 4957 scope.go:117] "RemoveContainer" containerID="87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.928023 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xjg58" event={"ID":"6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3","Type":"ContainerDied","Data":"d0b2c0b3692a16934278aa8feaea8497c9042680d05b2fc97b21325ef55092f6"} Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.928062 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0b2c0b3692a16934278aa8feaea8497c9042680d05b2fc97b21325ef55092f6" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.928116 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xjg58" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.932999 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44fz6\" (UniqueName: \"kubernetes.io/projected/97341258-bed0-4dda-9b61-bf6e9c93d375-kube-api-access-44fz6\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.933172 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.933272 4957 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.941382 4957 generic.go:334] "Generic (PLEG): container finished" podID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerID="2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e" exitCode=143 Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.941770 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="61d6eca8-07f2-473b-9e1e-34756cd9e33a" containerName="nova-scheduler-scheduler" containerID="cri-o://37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" gracePeriod=30 Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.941968 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0501ce8c-285f-4d42-abaa-1c7252324b0e","Type":"ContainerDied","Data":"2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e"} Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.958128 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.958883 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" containerName="nova-cell1-conductor-db-sync" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.958907 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" containerName="nova-cell1-conductor-db-sync" Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.958969 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-notification-agent" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.958980 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-notification-agent" Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.959028 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerName="dnsmasq-dns" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.959039 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerName="dnsmasq-dns" Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.959063 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" containerName="nova-manage" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.959073 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" containerName="nova-manage" Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.959118 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerName="init" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.959128 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerName="init" Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.959146 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-central-agent" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.959154 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-central-agent" Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.959218 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="proxy-httpd" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.959229 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="proxy-httpd" Dec 06 06:04:58 crc kubenswrapper[4957]: E1206 06:04:58.959249 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="sg-core" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.959256 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="sg-core" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.961150 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="sg-core" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.961201 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-central-agent" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.961220 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f840f7f6-e5f8-4307-910f-5c30200ae1fb" containerName="dnsmasq-dns" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.961265 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="proxy-httpd" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.961302 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" containerName="nova-cell1-conductor-db-sync" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.961342 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" containerName="ceilometer-notification-agent" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.961357 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" containerName="nova-manage" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.962508 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.966885 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.972123 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97341258-bed0-4dda-9b61-bf6e9c93d375" (UID: "97341258-bed0-4dda-9b61-bf6e9c93d375"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.984968 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 06:04:58 crc kubenswrapper[4957]: I1206 06:04:58.985123 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-config-data" (OuterVolumeSpecName: "config-data") pod "97341258-bed0-4dda-9b61-bf6e9c93d375" (UID: "97341258-bed0-4dda-9b61-bf6e9c93d375"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.010061 4957 scope.go:117] "RemoveContainer" containerID="a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.030262 4957 scope.go:117] "RemoveContainer" containerID="3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.034757 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.034857 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97341258-bed0-4dda-9b61-bf6e9c93d375-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.049018 4957 scope.go:117] "RemoveContainer" containerID="fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.067699 4957 scope.go:117] "RemoveContainer" containerID="87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c" Dec 06 06:04:59 crc kubenswrapper[4957]: E1206 06:04:59.068172 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c\": container with ID starting with 87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c not found: ID does not exist" containerID="87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.068216 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c"} err="failed to get container status \"87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c\": rpc error: code = NotFound desc = could not find container \"87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c\": container with ID starting with 87a3f053c20229c552a831f11d7d7ced9e52cb32aed65fb1b5098177ec07788c not found: ID does not exist" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.068248 4957 scope.go:117] "RemoveContainer" containerID="a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934" Dec 06 06:04:59 crc kubenswrapper[4957]: E1206 06:04:59.068516 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934\": container with ID starting with a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934 not found: ID does not exist" containerID="a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.068541 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934"} err="failed to get container status \"a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934\": rpc error: code = NotFound desc = could not find container \"a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934\": container with ID starting with a8adfc242c7f821688dd65e30b380dba43de35459857e2fe1ecd576de04da934 not found: ID does not exist" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.068555 4957 scope.go:117] "RemoveContainer" containerID="3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713" Dec 06 06:04:59 crc kubenswrapper[4957]: E1206 06:04:59.068787 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713\": container with ID starting with 3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713 not found: ID does not exist" containerID="3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.068823 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713"} err="failed to get container status \"3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713\": rpc error: code = NotFound desc = could not find container \"3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713\": container with ID starting with 3eeca4eb92a1b5b1d6bf65ce24501a0da7db4bccd064608ed00f82e0ded78713 not found: ID does not exist" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.068859 4957 scope.go:117] "RemoveContainer" containerID="fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a" Dec 06 06:04:59 crc kubenswrapper[4957]: E1206 06:04:59.069132 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a\": container with ID starting with fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a not found: ID does not exist" containerID="fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.069162 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a"} err="failed to get container status \"fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a\": rpc error: code = NotFound desc = could not find container \"fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a\": container with ID starting with fdf0ba880edba04fc252400ad988a8f1475c87c11e7116564fcb1526d0c5103a not found: ID does not exist" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.136138 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.136266 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.136293 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvnm6\" (UniqueName: \"kubernetes.io/projected/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-kube-api-access-gvnm6\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.238357 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.238632 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvnm6\" (UniqueName: \"kubernetes.io/projected/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-kube-api-access-gvnm6\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.238848 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.244537 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.244649 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.256579 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.265630 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvnm6\" (UniqueName: \"kubernetes.io/projected/6d950863-c9e9-4bc9-ae48-c8d1e63ad75c-kube-api-access-gvnm6\") pod \"nova-cell1-conductor-0\" (UID: \"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c\") " pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.281849 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.290618 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.293796 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.300287 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.300627 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.300795 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.301434 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.319305 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.441843 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-run-httpd\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.442174 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-log-httpd\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.442218 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.442253 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-config-data\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.442299 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-scripts\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.442338 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c9jp\" (UniqueName: \"kubernetes.io/projected/5858bf3b-46bf-4554-8954-930ef0e5c520-kube-api-access-8c9jp\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.442380 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.442411 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543653 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-scripts\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543715 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c9jp\" (UniqueName: \"kubernetes.io/projected/5858bf3b-46bf-4554-8954-930ef0e5c520-kube-api-access-8c9jp\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543772 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543802 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543866 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-run-httpd\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543892 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-log-httpd\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543906 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.543931 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-config-data\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.547426 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-log-httpd\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.547540 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-run-httpd\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.549132 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.549399 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-config-data\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.549762 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-scripts\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.554447 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.568132 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.568212 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c9jp\" (UniqueName: \"kubernetes.io/projected/5858bf3b-46bf-4554-8954-930ef0e5c520-kube-api-access-8c9jp\") pod \"ceilometer-0\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.613128 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:04:59 crc kubenswrapper[4957]: W1206 06:04:59.761026 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d950863_c9e9_4bc9_ae48_c8d1e63ad75c.slice/crio-a0c6fc238282a39685506e71f8db7f3726010694f251233d5bb798753fa0e2d4 WatchSource:0}: Error finding container a0c6fc238282a39685506e71f8db7f3726010694f251233d5bb798753fa0e2d4: Status 404 returned error can't find the container with id a0c6fc238282a39685506e71f8db7f3726010694f251233d5bb798753fa0e2d4 Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.784277 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 06:04:59 crc kubenswrapper[4957]: I1206 06:04:59.951720 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c","Type":"ContainerStarted","Data":"a0c6fc238282a39685506e71f8db7f3726010694f251233d5bb798753fa0e2d4"} Dec 06 06:05:00 crc kubenswrapper[4957]: I1206 06:05:00.094454 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:00 crc kubenswrapper[4957]: W1206 06:05:00.098569 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5858bf3b_46bf_4554_8954_930ef0e5c520.slice/crio-f257ff0ff60e93119a7db0f3b8314f4bb5bbab00bf3bfca0c884f82af507d162 WatchSource:0}: Error finding container f257ff0ff60e93119a7db0f3b8314f4bb5bbab00bf3bfca0c884f82af507d162: Status 404 returned error can't find the container with id f257ff0ff60e93119a7db0f3b8314f4bb5bbab00bf3bfca0c884f82af507d162 Dec 06 06:05:00 crc kubenswrapper[4957]: I1206 06:05:00.678826 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97341258-bed0-4dda-9b61-bf6e9c93d375" path="/var/lib/kubelet/pods/97341258-bed0-4dda-9b61-bf6e9c93d375/volumes" Dec 06 06:05:00 crc kubenswrapper[4957]: I1206 06:05:00.966954 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerStarted","Data":"32c7f95fe3499bee604afe9317411b10cce9de937287e8f59ac03aadbc673e05"} Dec 06 06:05:00 crc kubenswrapper[4957]: I1206 06:05:00.967020 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerStarted","Data":"f257ff0ff60e93119a7db0f3b8314f4bb5bbab00bf3bfca0c884f82af507d162"} Dec 06 06:05:00 crc kubenswrapper[4957]: I1206 06:05:00.968524 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6d950863-c9e9-4bc9-ae48-c8d1e63ad75c","Type":"ContainerStarted","Data":"1fec8cc3fa29cbc6b51fdd81922cd58186237cb95bc1ceab6edb4c7b7ef29ff3"} Dec 06 06:05:00 crc kubenswrapper[4957]: I1206 06:05:00.968948 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 06:05:00 crc kubenswrapper[4957]: I1206 06:05:00.986192 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.986164628 podStartE2EDuration="2.986164628s" podCreationTimestamp="2025-12-06 06:04:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:00.983142096 +0000 UTC m=+1540.633409738" watchObservedRunningTime="2025-12-06 06:05:00.986164628 +0000 UTC m=+1540.636432260" Dec 06 06:05:01 crc kubenswrapper[4957]: E1206 06:05:01.206399 4957 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:05:01 crc kubenswrapper[4957]: E1206 06:05:01.208970 4957 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:05:01 crc kubenswrapper[4957]: E1206 06:05:01.210405 4957 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:05:01 crc kubenswrapper[4957]: E1206 06:05:01.210507 4957 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="61d6eca8-07f2-473b-9e1e-34756cd9e33a" containerName="nova-scheduler-scheduler" Dec 06 06:05:01 crc kubenswrapper[4957]: I1206 06:05:01.979729 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerStarted","Data":"c9f24ce25b49ea89f386723e9796f6f1f1bc76469eaa3940e9dd51c9e3120308"} Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.760673 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.920375 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.927219 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft2l7\" (UniqueName: \"kubernetes.io/projected/61d6eca8-07f2-473b-9e1e-34756cd9e33a-kube-api-access-ft2l7\") pod \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.927271 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-combined-ca-bundle\") pod \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.927369 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-config-data\") pod \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\" (UID: \"61d6eca8-07f2-473b-9e1e-34756cd9e33a\") " Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.932436 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d6eca8-07f2-473b-9e1e-34756cd9e33a-kube-api-access-ft2l7" (OuterVolumeSpecName: "kube-api-access-ft2l7") pod "61d6eca8-07f2-473b-9e1e-34756cd9e33a" (UID: "61d6eca8-07f2-473b-9e1e-34756cd9e33a"). InnerVolumeSpecName "kube-api-access-ft2l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.955904 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-config-data" (OuterVolumeSpecName: "config-data") pod "61d6eca8-07f2-473b-9e1e-34756cd9e33a" (UID: "61d6eca8-07f2-473b-9e1e-34756cd9e33a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.965985 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61d6eca8-07f2-473b-9e1e-34756cd9e33a" (UID: "61d6eca8-07f2-473b-9e1e-34756cd9e33a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:02 crc kubenswrapper[4957]: I1206 06:05:02.996900 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerStarted","Data":"69030d56d92a36029f685a14df8c5ec40d2d3d005ce650e08db37f6885cb2024"} Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.005331 4957 generic.go:334] "Generic (PLEG): container finished" podID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerID="17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b" exitCode=0 Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.005389 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.005415 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0501ce8c-285f-4d42-abaa-1c7252324b0e","Type":"ContainerDied","Data":"17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b"} Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.005791 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0501ce8c-285f-4d42-abaa-1c7252324b0e","Type":"ContainerDied","Data":"a4627e08cbbd1768642639481054d496676167d66f4573b46e1b6fbf0e7547ec"} Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.005826 4957 scope.go:117] "RemoveContainer" containerID="17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.009575 4957 generic.go:334] "Generic (PLEG): container finished" podID="61d6eca8-07f2-473b-9e1e-34756cd9e33a" containerID="37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" exitCode=0 Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.009603 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61d6eca8-07f2-473b-9e1e-34756cd9e33a","Type":"ContainerDied","Data":"37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e"} Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.009622 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61d6eca8-07f2-473b-9e1e-34756cd9e33a","Type":"ContainerDied","Data":"481ff8e4f9f210afe0d83cda1282458e217db4e6d0223be09be3603dc2f9ff64"} Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.009664 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.029435 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0501ce8c-285f-4d42-abaa-1c7252324b0e-logs\") pod \"0501ce8c-285f-4d42-abaa-1c7252324b0e\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.029767 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pb5f\" (UniqueName: \"kubernetes.io/projected/0501ce8c-285f-4d42-abaa-1c7252324b0e-kube-api-access-4pb5f\") pod \"0501ce8c-285f-4d42-abaa-1c7252324b0e\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.029924 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0501ce8c-285f-4d42-abaa-1c7252324b0e-logs" (OuterVolumeSpecName: "logs") pod "0501ce8c-285f-4d42-abaa-1c7252324b0e" (UID: "0501ce8c-285f-4d42-abaa-1c7252324b0e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.030307 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-config-data\") pod \"0501ce8c-285f-4d42-abaa-1c7252324b0e\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.036400 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-combined-ca-bundle\") pod \"0501ce8c-285f-4d42-abaa-1c7252324b0e\" (UID: \"0501ce8c-285f-4d42-abaa-1c7252324b0e\") " Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.037148 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft2l7\" (UniqueName: \"kubernetes.io/projected/61d6eca8-07f2-473b-9e1e-34756cd9e33a-kube-api-access-ft2l7\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.037168 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.037180 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61d6eca8-07f2-473b-9e1e-34756cd9e33a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.037192 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0501ce8c-285f-4d42-abaa-1c7252324b0e-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.041510 4957 scope.go:117] "RemoveContainer" containerID="2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.049284 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0501ce8c-285f-4d42-abaa-1c7252324b0e-kube-api-access-4pb5f" (OuterVolumeSpecName: "kube-api-access-4pb5f") pod "0501ce8c-285f-4d42-abaa-1c7252324b0e" (UID: "0501ce8c-285f-4d42-abaa-1c7252324b0e"). InnerVolumeSpecName "kube-api-access-4pb5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.062407 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.064602 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0501ce8c-285f-4d42-abaa-1c7252324b0e" (UID: "0501ce8c-285f-4d42-abaa-1c7252324b0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.095062 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.101843 4957 scope.go:117] "RemoveContainer" containerID="17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b" Dec 06 06:05:03 crc kubenswrapper[4957]: E1206 06:05:03.104476 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b\": container with ID starting with 17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b not found: ID does not exist" containerID="17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.104507 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b"} err="failed to get container status \"17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b\": rpc error: code = NotFound desc = could not find container \"17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b\": container with ID starting with 17969a59742e38e1e06bff5610e9a92a237f6f68e0b99babb8ade20c4fef334b not found: ID does not exist" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.104529 4957 scope.go:117] "RemoveContainer" containerID="2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.107546 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: E1206 06:05:03.107989 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-api" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108005 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-api" Dec 06 06:05:03 crc kubenswrapper[4957]: E1206 06:05:03.108012 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d6eca8-07f2-473b-9e1e-34756cd9e33a" containerName="nova-scheduler-scheduler" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108020 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d6eca8-07f2-473b-9e1e-34756cd9e33a" containerName="nova-scheduler-scheduler" Dec 06 06:05:03 crc kubenswrapper[4957]: E1206 06:05:03.108050 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-log" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108056 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-log" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108217 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d6eca8-07f2-473b-9e1e-34756cd9e33a" containerName="nova-scheduler-scheduler" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108229 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-log" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108252 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" containerName="nova-api-api" Dec 06 06:05:03 crc kubenswrapper[4957]: E1206 06:05:03.108293 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e\": container with ID starting with 2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e not found: ID does not exist" containerID="2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108319 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e"} err="failed to get container status \"2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e\": rpc error: code = NotFound desc = could not find container \"2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e\": container with ID starting with 2fdd24f324abe213fd5437512f530438bc2c7babc932084e9ab7e11a1931354e not found: ID does not exist" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.108332 4957 scope.go:117] "RemoveContainer" containerID="37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.109003 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.109978 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-config-data" (OuterVolumeSpecName: "config-data") pod "0501ce8c-285f-4d42-abaa-1c7252324b0e" (UID: "0501ce8c-285f-4d42-abaa-1c7252324b0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.113465 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.115145 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.138515 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pb5f\" (UniqueName: \"kubernetes.io/projected/0501ce8c-285f-4d42-abaa-1c7252324b0e-kube-api-access-4pb5f\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.138547 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.138557 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0501ce8c-285f-4d42-abaa-1c7252324b0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.144725 4957 scope.go:117] "RemoveContainer" containerID="37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" Dec 06 06:05:03 crc kubenswrapper[4957]: E1206 06:05:03.145239 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e\": container with ID starting with 37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e not found: ID does not exist" containerID="37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.145270 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e"} err="failed to get container status \"37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e\": rpc error: code = NotFound desc = could not find container \"37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e\": container with ID starting with 37640ab067cd3330d72468f81af65e86b7c81a51345420ecb37e0ac2ec6d5d6e not found: ID does not exist" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.239634 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-config-data\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.239980 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rs2h\" (UniqueName: \"kubernetes.io/projected/b3b61136-8d49-4883-b52c-f370b552d208-kube-api-access-5rs2h\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.240067 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.336253 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.341071 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.341193 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-config-data\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.341232 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rs2h\" (UniqueName: \"kubernetes.io/projected/b3b61136-8d49-4883-b52c-f370b552d208-kube-api-access-5rs2h\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.346687 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-config-data\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.346709 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.350651 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.366560 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rs2h\" (UniqueName: \"kubernetes.io/projected/b3b61136-8d49-4883-b52c-f370b552d208-kube-api-access-5rs2h\") pod \"nova-scheduler-0\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.368494 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.370105 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.375219 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.380991 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.431928 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.442420 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjlv9\" (UniqueName: \"kubernetes.io/projected/211de15b-f503-4bc0-afbc-4ffb34813fe4-kube-api-access-kjlv9\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.442457 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.442491 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-config-data\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.442636 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/211de15b-f503-4bc0-afbc-4ffb34813fe4-logs\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.559487 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjlv9\" (UniqueName: \"kubernetes.io/projected/211de15b-f503-4bc0-afbc-4ffb34813fe4-kube-api-access-kjlv9\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.559925 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.559981 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-config-data\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.560029 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/211de15b-f503-4bc0-afbc-4ffb34813fe4-logs\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.560658 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/211de15b-f503-4bc0-afbc-4ffb34813fe4-logs\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.566638 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.575489 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-config-data\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.598455 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjlv9\" (UniqueName: \"kubernetes.io/projected/211de15b-f503-4bc0-afbc-4ffb34813fe4-kube-api-access-kjlv9\") pod \"nova-api-0\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.690181 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:03 crc kubenswrapper[4957]: I1206 06:05:03.906376 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.029367 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3b61136-8d49-4883-b52c-f370b552d208","Type":"ContainerStarted","Data":"11606f6c949cdc6f9c52fdb27c32406f82b550638b6dd55bc1380a5e5b4ae524"} Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.034954 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerStarted","Data":"f6adbbc76f940e8f02404f5edfc473d56f0032014dd24f5287a76f2d9c2040a0"} Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.036027 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.061972 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.602272946 podStartE2EDuration="5.061950219s" podCreationTimestamp="2025-12-06 06:04:59 +0000 UTC" firstStartedPulling="2025-12-06 06:05:00.101860363 +0000 UTC m=+1539.752127995" lastFinishedPulling="2025-12-06 06:05:03.561537636 +0000 UTC m=+1543.211805268" observedRunningTime="2025-12-06 06:05:04.054543889 +0000 UTC m=+1543.704811541" watchObservedRunningTime="2025-12-06 06:05:04.061950219 +0000 UTC m=+1543.712217861" Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.200158 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 06:05:04 crc kubenswrapper[4957]: W1206 06:05:04.211602 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod211de15b_f503_4bc0_afbc_4ffb34813fe4.slice/crio-e0a3ff7de91abbd5846cf8b5466da3c9ecabc1294e9855bdd25758e8623e4bb4 WatchSource:0}: Error finding container e0a3ff7de91abbd5846cf8b5466da3c9ecabc1294e9855bdd25758e8623e4bb4: Status 404 returned error can't find the container with id e0a3ff7de91abbd5846cf8b5466da3c9ecabc1294e9855bdd25758e8623e4bb4 Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.213611 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.365429 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.672490 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0501ce8c-285f-4d42-abaa-1c7252324b0e" path="/var/lib/kubelet/pods/0501ce8c-285f-4d42-abaa-1c7252324b0e/volumes" Dec 06 06:05:04 crc kubenswrapper[4957]: I1206 06:05:04.673067 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d6eca8-07f2-473b-9e1e-34756cd9e33a" path="/var/lib/kubelet/pods/61d6eca8-07f2-473b-9e1e-34756cd9e33a/volumes" Dec 06 06:05:05 crc kubenswrapper[4957]: I1206 06:05:05.048962 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"211de15b-f503-4bc0-afbc-4ffb34813fe4","Type":"ContainerStarted","Data":"ab92b41a089bdbc3f9affc910fb43c1381c6e181e5fb1629325b2a09a2adf2cd"} Dec 06 06:05:05 crc kubenswrapper[4957]: I1206 06:05:05.049037 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"211de15b-f503-4bc0-afbc-4ffb34813fe4","Type":"ContainerStarted","Data":"af10de0f59cc5189c61a329366ab6b05ff27dfedd1b2c642a40e66fee534da4f"} Dec 06 06:05:05 crc kubenswrapper[4957]: I1206 06:05:05.049082 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"211de15b-f503-4bc0-afbc-4ffb34813fe4","Type":"ContainerStarted","Data":"e0a3ff7de91abbd5846cf8b5466da3c9ecabc1294e9855bdd25758e8623e4bb4"} Dec 06 06:05:05 crc kubenswrapper[4957]: I1206 06:05:05.050676 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3b61136-8d49-4883-b52c-f370b552d208","Type":"ContainerStarted","Data":"e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5"} Dec 06 06:05:05 crc kubenswrapper[4957]: I1206 06:05:05.067858 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.067822012 podStartE2EDuration="2.067822012s" podCreationTimestamp="2025-12-06 06:05:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:05.065718325 +0000 UTC m=+1544.715985967" watchObservedRunningTime="2025-12-06 06:05:05.067822012 +0000 UTC m=+1544.718089644" Dec 06 06:05:05 crc kubenswrapper[4957]: I1206 06:05:05.084630 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.084607856 podStartE2EDuration="2.084607856s" podCreationTimestamp="2025-12-06 06:05:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:05.082800927 +0000 UTC m=+1544.733068579" watchObservedRunningTime="2025-12-06 06:05:05.084607856 +0000 UTC m=+1544.734875488" Dec 06 06:05:08 crc kubenswrapper[4957]: I1206 06:05:08.432506 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 06:05:10 crc kubenswrapper[4957]: I1206 06:05:10.180411 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:05:10 crc kubenswrapper[4957]: I1206 06:05:10.181210 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:05:13 crc kubenswrapper[4957]: I1206 06:05:13.433036 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 06:05:13 crc kubenswrapper[4957]: I1206 06:05:13.468796 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 06:05:13 crc kubenswrapper[4957]: I1206 06:05:13.690515 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:05:13 crc kubenswrapper[4957]: I1206 06:05:13.691291 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:05:14 crc kubenswrapper[4957]: I1206 06:05:14.181866 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 06:05:14 crc kubenswrapper[4957]: I1206 06:05:14.773122 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:05:14 crc kubenswrapper[4957]: I1206 06:05:14.773439 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.217797 4957 generic.go:334] "Generic (PLEG): container finished" podID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerID="4c8b6674d15d7b91045bb18c121ac04a1270deda62a69aab95c9d178013e71b0" exitCode=137 Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.218395 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c757adb-1ed5-40b7-a923-79b89d269f1e","Type":"ContainerDied","Data":"4c8b6674d15d7b91045bb18c121ac04a1270deda62a69aab95c9d178013e71b0"} Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.218431 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c757adb-1ed5-40b7-a923-79b89d269f1e","Type":"ContainerDied","Data":"7970c0ae2e19fe0826a397c6f2c4b14d52a964340e63a5ea8cf24d358178994c"} Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.218446 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7970c0ae2e19fe0826a397c6f2c4b14d52a964340e63a5ea8cf24d358178994c" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.220339 4957 generic.go:334] "Generic (PLEG): container finished" podID="82cd2252-346e-4456-b059-193abd84d10d" containerID="c6cb89a94d7d6e6ce03f464bc9ac4b7a8ed6e181d3c005801e90785499667870" exitCode=137 Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.220367 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82cd2252-346e-4456-b059-193abd84d10d","Type":"ContainerDied","Data":"c6cb89a94d7d6e6ce03f464bc9ac4b7a8ed6e181d3c005801e90785499667870"} Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.220382 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82cd2252-346e-4456-b059-193abd84d10d","Type":"ContainerDied","Data":"68c25519ceef1a60ed88408e219dcdd1031410d246ea0547a3a96602f80759d8"} Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.220392 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68c25519ceef1a60ed88408e219dcdd1031410d246ea0547a3a96602f80759d8" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.244901 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.248550 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.265697 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz4fd\" (UniqueName: \"kubernetes.io/projected/5c757adb-1ed5-40b7-a923-79b89d269f1e-kube-api-access-lz4fd\") pod \"5c757adb-1ed5-40b7-a923-79b89d269f1e\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.265791 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c757adb-1ed5-40b7-a923-79b89d269f1e-logs\") pod \"5c757adb-1ed5-40b7-a923-79b89d269f1e\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.265855 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-config-data\") pod \"5c757adb-1ed5-40b7-a923-79b89d269f1e\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.266375 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c757adb-1ed5-40b7-a923-79b89d269f1e-logs" (OuterVolumeSpecName: "logs") pod "5c757adb-1ed5-40b7-a923-79b89d269f1e" (UID: "5c757adb-1ed5-40b7-a923-79b89d269f1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.265904 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-config-data\") pod \"82cd2252-346e-4456-b059-193abd84d10d\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.269545 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-combined-ca-bundle\") pod \"5c757adb-1ed5-40b7-a923-79b89d269f1e\" (UID: \"5c757adb-1ed5-40b7-a923-79b89d269f1e\") " Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.269591 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh9cb\" (UniqueName: \"kubernetes.io/projected/82cd2252-346e-4456-b059-193abd84d10d-kube-api-access-zh9cb\") pod \"82cd2252-346e-4456-b059-193abd84d10d\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.270105 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-combined-ca-bundle\") pod \"82cd2252-346e-4456-b059-193abd84d10d\" (UID: \"82cd2252-346e-4456-b059-193abd84d10d\") " Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.271237 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c757adb-1ed5-40b7-a923-79b89d269f1e-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.273304 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82cd2252-346e-4456-b059-193abd84d10d-kube-api-access-zh9cb" (OuterVolumeSpecName: "kube-api-access-zh9cb") pod "82cd2252-346e-4456-b059-193abd84d10d" (UID: "82cd2252-346e-4456-b059-193abd84d10d"). InnerVolumeSpecName "kube-api-access-zh9cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.274028 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c757adb-1ed5-40b7-a923-79b89d269f1e-kube-api-access-lz4fd" (OuterVolumeSpecName: "kube-api-access-lz4fd") pod "5c757adb-1ed5-40b7-a923-79b89d269f1e" (UID: "5c757adb-1ed5-40b7-a923-79b89d269f1e"). InnerVolumeSpecName "kube-api-access-lz4fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.299998 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c757adb-1ed5-40b7-a923-79b89d269f1e" (UID: "5c757adb-1ed5-40b7-a923-79b89d269f1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.313263 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82cd2252-346e-4456-b059-193abd84d10d" (UID: "82cd2252-346e-4456-b059-193abd84d10d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.318508 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-config-data" (OuterVolumeSpecName: "config-data") pod "5c757adb-1ed5-40b7-a923-79b89d269f1e" (UID: "5c757adb-1ed5-40b7-a923-79b89d269f1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.328745 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-config-data" (OuterVolumeSpecName: "config-data") pod "82cd2252-346e-4456-b059-193abd84d10d" (UID: "82cd2252-346e-4456-b059-193abd84d10d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.379355 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.379405 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz4fd\" (UniqueName: \"kubernetes.io/projected/5c757adb-1ed5-40b7-a923-79b89d269f1e-kube-api-access-lz4fd\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.379421 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.379433 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82cd2252-346e-4456-b059-193abd84d10d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.379448 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c757adb-1ed5-40b7-a923-79b89d269f1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:21 crc kubenswrapper[4957]: I1206 06:05:21.379460 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh9cb\" (UniqueName: \"kubernetes.io/projected/82cd2252-346e-4456-b059-193abd84d10d-kube-api-access-zh9cb\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.229263 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.229440 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.268934 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.284873 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.310242 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.357940 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.375991 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: E1206 06:05:22.376704 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82cd2252-346e-4456-b059-193abd84d10d" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.376731 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="82cd2252-346e-4456-b059-193abd84d10d" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 06:05:22 crc kubenswrapper[4957]: E1206 06:05:22.376769 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-log" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.376781 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-log" Dec 06 06:05:22 crc kubenswrapper[4957]: E1206 06:05:22.376807 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-metadata" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.376817 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-metadata" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.377044 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="82cd2252-346e-4456-b059-193abd84d10d" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.377066 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-metadata" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.377101 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" containerName="nova-metadata-log" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.378372 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.380649 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.382275 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.394256 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.398923 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7dc285-d8d8-4d28-9335-9f0abaa31738-logs\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.399064 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.399095 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.399172 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-config-data\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.399201 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkgl8\" (UniqueName: \"kubernetes.io/projected/af7dc285-d8d8-4d28-9335-9f0abaa31738-kube-api-access-fkgl8\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.406852 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.408125 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.413799 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.413848 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.415715 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.416278 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.500742 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.500878 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7dc285-d8d8-4d28-9335-9f0abaa31738-logs\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.501007 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.501055 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.501083 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75m54\" (UniqueName: \"kubernetes.io/projected/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-kube-api-access-75m54\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.501171 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.501284 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-config-data\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.501395 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkgl8\" (UniqueName: \"kubernetes.io/projected/af7dc285-d8d8-4d28-9335-9f0abaa31738-kube-api-access-fkgl8\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.501462 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7dc285-d8d8-4d28-9335-9f0abaa31738-logs\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.502295 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.502351 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.506712 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.506752 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-config-data\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.506731 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.524472 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkgl8\" (UniqueName: \"kubernetes.io/projected/af7dc285-d8d8-4d28-9335-9f0abaa31738-kube-api-access-fkgl8\") pod \"nova-metadata-0\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.604353 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.604438 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.604466 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.604534 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.604640 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75m54\" (UniqueName: \"kubernetes.io/projected/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-kube-api-access-75m54\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.607919 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.607984 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.609790 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.610007 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.620121 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75m54\" (UniqueName: \"kubernetes.io/projected/ecb88cbc-d465-49e9-9aaf-31aa46e18ae2-kube-api-access-75m54\") pod \"nova-cell1-novncproxy-0\" (UID: \"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.680371 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c757adb-1ed5-40b7-a923-79b89d269f1e" path="/var/lib/kubelet/pods/5c757adb-1ed5-40b7-a923-79b89d269f1e/volumes" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.681325 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82cd2252-346e-4456-b059-193abd84d10d" path="/var/lib/kubelet/pods/82cd2252-346e-4456-b059-193abd84d10d/volumes" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.696026 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:22 crc kubenswrapper[4957]: I1206 06:05:22.726045 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.165297 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.215007 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 06:05:23 crc kubenswrapper[4957]: W1206 06:05:23.216431 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecb88cbc_d465_49e9_9aaf_31aa46e18ae2.slice/crio-49ee48bcfe708d889d0229da772def2c1586d73cc112337e68732c3f3fa60728 WatchSource:0}: Error finding container 49ee48bcfe708d889d0229da772def2c1586d73cc112337e68732c3f3fa60728: Status 404 returned error can't find the container with id 49ee48bcfe708d889d0229da772def2c1586d73cc112337e68732c3f3fa60728 Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.239337 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af7dc285-d8d8-4d28-9335-9f0abaa31738","Type":"ContainerStarted","Data":"329cdff5540f679ae3a7368e72c4b9c1965f07a122d6d800c2a57a268f9711a7"} Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.240734 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2","Type":"ContainerStarted","Data":"49ee48bcfe708d889d0229da772def2c1586d73cc112337e68732c3f3fa60728"} Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.693913 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.696041 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.700510 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:05:23 crc kubenswrapper[4957]: I1206 06:05:23.704152 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.251079 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ecb88cbc-d465-49e9-9aaf-31aa46e18ae2","Type":"ContainerStarted","Data":"a54e6188f647a6f45bfa97eaf67271c4730eb207a6814c4276c72d90a9a017bb"} Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.260212 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af7dc285-d8d8-4d28-9335-9f0abaa31738","Type":"ContainerStarted","Data":"183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36"} Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.260248 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af7dc285-d8d8-4d28-9335-9f0abaa31738","Type":"ContainerStarted","Data":"196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9"} Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.260644 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.263731 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.283709 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.283682915 podStartE2EDuration="2.283682915s" podCreationTimestamp="2025-12-06 06:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:24.271555114 +0000 UTC m=+1563.921822776" watchObservedRunningTime="2025-12-06 06:05:24.283682915 +0000 UTC m=+1563.933950577" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.294764 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.294746367 podStartE2EDuration="2.294746367s" podCreationTimestamp="2025-12-06 06:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:24.293227936 +0000 UTC m=+1563.943495578" watchObservedRunningTime="2025-12-06 06:05:24.294746367 +0000 UTC m=+1563.945013999" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.472167 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vlnfn"] Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.473775 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.500315 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vlnfn"] Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.569068 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hxhx\" (UniqueName: \"kubernetes.io/projected/a2ac66e5-0300-49e5-9193-466fd0520dd6-kube-api-access-6hxhx\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.569232 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.569274 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.569313 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-config\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.569366 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.569399 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.670732 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.670778 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.670846 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hxhx\" (UniqueName: \"kubernetes.io/projected/a2ac66e5-0300-49e5-9193-466fd0520dd6-kube-api-access-6hxhx\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.670965 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.670992 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.671024 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-config\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.671897 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.671926 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.672015 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-config\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.672401 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.672503 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.689917 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hxhx\" (UniqueName: \"kubernetes.io/projected/a2ac66e5-0300-49e5-9193-466fd0520dd6-kube-api-access-6hxhx\") pod \"dnsmasq-dns-89c5cd4d5-vlnfn\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:24 crc kubenswrapper[4957]: I1206 06:05:24.817277 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:25 crc kubenswrapper[4957]: I1206 06:05:25.358965 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vlnfn"] Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.275107 4957 generic.go:334] "Generic (PLEG): container finished" podID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerID="b6eb501e08ec6789d09c4c9926e664db4b88841e68e31a4d04c3a3f5ff56fd4d" exitCode=0 Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.276341 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" event={"ID":"a2ac66e5-0300-49e5-9193-466fd0520dd6","Type":"ContainerDied","Data":"b6eb501e08ec6789d09c4c9926e664db4b88841e68e31a4d04c3a3f5ff56fd4d"} Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.276374 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" event={"ID":"a2ac66e5-0300-49e5-9193-466fd0520dd6","Type":"ContainerStarted","Data":"66f91fbf68a73a7ab250253d38d1520b2a1f86c240f3048b23416c6fa5acdaeb"} Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.612483 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.612976 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-central-agent" containerID="cri-o://32c7f95fe3499bee604afe9317411b10cce9de937287e8f59ac03aadbc673e05" gracePeriod=30 Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.613057 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="proxy-httpd" containerID="cri-o://f6adbbc76f940e8f02404f5edfc473d56f0032014dd24f5287a76f2d9c2040a0" gracePeriod=30 Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.613067 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-notification-agent" containerID="cri-o://c9f24ce25b49ea89f386723e9796f6f1f1bc76469eaa3940e9dd51c9e3120308" gracePeriod=30 Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.613103 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="sg-core" containerID="cri-o://69030d56d92a36029f685a14df8c5ec40d2d3d005ce650e08db37f6885cb2024" gracePeriod=30 Dec 06 06:05:26 crc kubenswrapper[4957]: I1206 06:05:26.623201 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.194:3000/\": EOF" Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.287963 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" event={"ID":"a2ac66e5-0300-49e5-9193-466fd0520dd6","Type":"ContainerStarted","Data":"4daee8ec0c6f6b263e156056c46796dfe9e76dc423527041ed2301d6695cac09"} Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.288986 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.291117 4957 generic.go:334] "Generic (PLEG): container finished" podID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerID="f6adbbc76f940e8f02404f5edfc473d56f0032014dd24f5287a76f2d9c2040a0" exitCode=0 Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.291143 4957 generic.go:334] "Generic (PLEG): container finished" podID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerID="69030d56d92a36029f685a14df8c5ec40d2d3d005ce650e08db37f6885cb2024" exitCode=2 Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.291150 4957 generic.go:334] "Generic (PLEG): container finished" podID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerID="32c7f95fe3499bee604afe9317411b10cce9de937287e8f59ac03aadbc673e05" exitCode=0 Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.291166 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerDied","Data":"f6adbbc76f940e8f02404f5edfc473d56f0032014dd24f5287a76f2d9c2040a0"} Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.291187 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerDied","Data":"69030d56d92a36029f685a14df8c5ec40d2d3d005ce650e08db37f6885cb2024"} Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.291198 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerDied","Data":"32c7f95fe3499bee604afe9317411b10cce9de937287e8f59ac03aadbc673e05"} Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.331105 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" podStartSLOduration=3.331082281 podStartE2EDuration="3.331082281s" podCreationTimestamp="2025-12-06 06:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:27.317604833 +0000 UTC m=+1566.967872525" watchObservedRunningTime="2025-12-06 06:05:27.331082281 +0000 UTC m=+1566.981349933" Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.696867 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.696974 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.727222 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.955303 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.956327 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-log" containerID="cri-o://af10de0f59cc5189c61a329366ab6b05ff27dfedd1b2c642a40e66fee534da4f" gracePeriod=30 Dec 06 06:05:27 crc kubenswrapper[4957]: I1206 06:05:27.956381 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-api" containerID="cri-o://ab92b41a089bdbc3f9affc910fb43c1381c6e181e5fb1629325b2a09a2adf2cd" gracePeriod=30 Dec 06 06:05:28 crc kubenswrapper[4957]: I1206 06:05:28.302977 4957 generic.go:334] "Generic (PLEG): container finished" podID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerID="af10de0f59cc5189c61a329366ab6b05ff27dfedd1b2c642a40e66fee534da4f" exitCode=143 Dec 06 06:05:28 crc kubenswrapper[4957]: I1206 06:05:28.303791 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"211de15b-f503-4bc0-afbc-4ffb34813fe4","Type":"ContainerDied","Data":"af10de0f59cc5189c61a329366ab6b05ff27dfedd1b2c642a40e66fee534da4f"} Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.315455 4957 generic.go:334] "Generic (PLEG): container finished" podID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerID="c9f24ce25b49ea89f386723e9796f6f1f1bc76469eaa3940e9dd51c9e3120308" exitCode=0 Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.315502 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerDied","Data":"c9f24ce25b49ea89f386723e9796f6f1f1bc76469eaa3940e9dd51c9e3120308"} Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.572301 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621184 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-log-httpd\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621274 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-scripts\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621367 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-combined-ca-bundle\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621432 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-config-data\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621487 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-sg-core-conf-yaml\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621587 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c9jp\" (UniqueName: \"kubernetes.io/projected/5858bf3b-46bf-4554-8954-930ef0e5c520-kube-api-access-8c9jp\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621662 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621741 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-run-httpd\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.621879 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-ceilometer-tls-certs\") pod \"5858bf3b-46bf-4554-8954-930ef0e5c520\" (UID: \"5858bf3b-46bf-4554-8954-930ef0e5c520\") " Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.622591 4957 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.626215 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.629748 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-scripts" (OuterVolumeSpecName: "scripts") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.632006 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5858bf3b-46bf-4554-8954-930ef0e5c520-kube-api-access-8c9jp" (OuterVolumeSpecName: "kube-api-access-8c9jp") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "kube-api-access-8c9jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.673971 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.725944 4957 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5858bf3b-46bf-4554-8954-930ef0e5c520-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.725981 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.725991 4957 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.726025 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c9jp\" (UniqueName: \"kubernetes.io/projected/5858bf3b-46bf-4554-8954-930ef0e5c520-kube-api-access-8c9jp\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.730374 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.751790 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.770180 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-config-data" (OuterVolumeSpecName: "config-data") pod "5858bf3b-46bf-4554-8954-930ef0e5c520" (UID: "5858bf3b-46bf-4554-8954-930ef0e5c520"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.827585 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.827621 4957 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:29 crc kubenswrapper[4957]: I1206 06:05:29.827630 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5858bf3b-46bf-4554-8954-930ef0e5c520-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.330290 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5858bf3b-46bf-4554-8954-930ef0e5c520","Type":"ContainerDied","Data":"f257ff0ff60e93119a7db0f3b8314f4bb5bbab00bf3bfca0c884f82af507d162"} Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.330491 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.331286 4957 scope.go:117] "RemoveContainer" containerID="f6adbbc76f940e8f02404f5edfc473d56f0032014dd24f5287a76f2d9c2040a0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.363097 4957 scope.go:117] "RemoveContainer" containerID="69030d56d92a36029f685a14df8c5ec40d2d3d005ce650e08db37f6885cb2024" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.384981 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.392071 4957 scope.go:117] "RemoveContainer" containerID="c9f24ce25b49ea89f386723e9796f6f1f1bc76469eaa3940e9dd51c9e3120308" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.401714 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.419605 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:30 crc kubenswrapper[4957]: E1206 06:05:30.420090 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="proxy-httpd" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420102 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="proxy-httpd" Dec 06 06:05:30 crc kubenswrapper[4957]: E1206 06:05:30.420129 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-central-agent" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420136 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-central-agent" Dec 06 06:05:30 crc kubenswrapper[4957]: E1206 06:05:30.420151 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-notification-agent" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420158 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-notification-agent" Dec 06 06:05:30 crc kubenswrapper[4957]: E1206 06:05:30.420170 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="sg-core" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420176 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="sg-core" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420358 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="sg-core" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420378 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="proxy-httpd" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420387 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-notification-agent" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.420396 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" containerName="ceilometer-central-agent" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.422096 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.422173 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.453803 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-config-data\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.453912 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.453933 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.453974 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ww6c\" (UniqueName: \"kubernetes.io/projected/30a9c22f-d5a5-4794-9f6b-70a16656726a-kube-api-access-9ww6c\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.454033 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.454064 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-run-httpd\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.454084 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-log-httpd\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.454201 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-scripts\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.454869 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.455140 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.455257 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.479960 4957 scope.go:117] "RemoveContainer" containerID="32c7f95fe3499bee604afe9317411b10cce9de937287e8f59ac03aadbc673e05" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556502 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556559 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-run-httpd\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556583 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-log-httpd\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556643 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-scripts\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556708 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-config-data\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556767 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556795 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.556856 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ww6c\" (UniqueName: \"kubernetes.io/projected/30a9c22f-d5a5-4794-9f6b-70a16656726a-kube-api-access-9ww6c\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.557186 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-run-httpd\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.557417 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-log-httpd\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.561903 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-scripts\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.562443 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.562940 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.563286 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-config-data\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.563409 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.574070 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ww6c\" (UniqueName: \"kubernetes.io/projected/30a9c22f-d5a5-4794-9f6b-70a16656726a-kube-api-access-9ww6c\") pod \"ceilometer-0\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.674678 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5858bf3b-46bf-4554-8954-930ef0e5c520" path="/var/lib/kubelet/pods/5858bf3b-46bf-4554-8954-930ef0e5c520/volumes" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.772085 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:05:30 crc kubenswrapper[4957]: I1206 06:05:30.815420 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.325444 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.343978 4957 generic.go:334] "Generic (PLEG): container finished" podID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerID="ab92b41a089bdbc3f9affc910fb43c1381c6e181e5fb1629325b2a09a2adf2cd" exitCode=0 Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.344080 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"211de15b-f503-4bc0-afbc-4ffb34813fe4","Type":"ContainerDied","Data":"ab92b41a089bdbc3f9affc910fb43c1381c6e181e5fb1629325b2a09a2adf2cd"} Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.507995 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.575511 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-config-data\") pod \"211de15b-f503-4bc0-afbc-4ffb34813fe4\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.575555 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjlv9\" (UniqueName: \"kubernetes.io/projected/211de15b-f503-4bc0-afbc-4ffb34813fe4-kube-api-access-kjlv9\") pod \"211de15b-f503-4bc0-afbc-4ffb34813fe4\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.575625 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-combined-ca-bundle\") pod \"211de15b-f503-4bc0-afbc-4ffb34813fe4\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.575645 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/211de15b-f503-4bc0-afbc-4ffb34813fe4-logs\") pod \"211de15b-f503-4bc0-afbc-4ffb34813fe4\" (UID: \"211de15b-f503-4bc0-afbc-4ffb34813fe4\") " Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.576734 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/211de15b-f503-4bc0-afbc-4ffb34813fe4-logs" (OuterVolumeSpecName: "logs") pod "211de15b-f503-4bc0-afbc-4ffb34813fe4" (UID: "211de15b-f503-4bc0-afbc-4ffb34813fe4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.589079 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/211de15b-f503-4bc0-afbc-4ffb34813fe4-kube-api-access-kjlv9" (OuterVolumeSpecName: "kube-api-access-kjlv9") pod "211de15b-f503-4bc0-afbc-4ffb34813fe4" (UID: "211de15b-f503-4bc0-afbc-4ffb34813fe4"). InnerVolumeSpecName "kube-api-access-kjlv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.609871 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-config-data" (OuterVolumeSpecName: "config-data") pod "211de15b-f503-4bc0-afbc-4ffb34813fe4" (UID: "211de15b-f503-4bc0-afbc-4ffb34813fe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.620277 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "211de15b-f503-4bc0-afbc-4ffb34813fe4" (UID: "211de15b-f503-4bc0-afbc-4ffb34813fe4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.678863 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.679047 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/211de15b-f503-4bc0-afbc-4ffb34813fe4-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.679335 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211de15b-f503-4bc0-afbc-4ffb34813fe4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:31 crc kubenswrapper[4957]: I1206 06:05:31.679348 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjlv9\" (UniqueName: \"kubernetes.io/projected/211de15b-f503-4bc0-afbc-4ffb34813fe4-kube-api-access-kjlv9\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.359142 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerStarted","Data":"e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc"} Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.359716 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerStarted","Data":"f6384d890b53464cca00650f32f0b2e1d51800a4e4593e074f1b67cec4ad15aa"} Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.361938 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"211de15b-f503-4bc0-afbc-4ffb34813fe4","Type":"ContainerDied","Data":"e0a3ff7de91abbd5846cf8b5466da3c9ecabc1294e9855bdd25758e8623e4bb4"} Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.361989 4957 scope.go:117] "RemoveContainer" containerID="ab92b41a089bdbc3f9affc910fb43c1381c6e181e5fb1629325b2a09a2adf2cd" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.362105 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.398606 4957 scope.go:117] "RemoveContainer" containerID="af10de0f59cc5189c61a329366ab6b05ff27dfedd1b2c642a40e66fee534da4f" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.407020 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.422627 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.450351 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:32 crc kubenswrapper[4957]: E1206 06:05:32.450870 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-log" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.450893 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-log" Dec 06 06:05:32 crc kubenswrapper[4957]: E1206 06:05:32.450925 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-api" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.450935 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-api" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.451200 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-api" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.451231 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" containerName="nova-api-log" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.452504 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.454627 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.455131 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.455295 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.461946 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.605763 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b324875-8f1a-41f5-b9e7-63c004cebba3-logs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.605877 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.605909 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l8zl\" (UniqueName: \"kubernetes.io/projected/1b324875-8f1a-41f5-b9e7-63c004cebba3-kube-api-access-9l8zl\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.605930 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.606034 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-config-data\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.606102 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.694068 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="211de15b-f503-4bc0-afbc-4ffb34813fe4" path="/var/lib/kubelet/pods/211de15b-f503-4bc0-afbc-4ffb34813fe4/volumes" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.698658 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.698708 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.707164 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-config-data\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.707252 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.707296 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b324875-8f1a-41f5-b9e7-63c004cebba3-logs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.707334 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.707363 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.707378 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l8zl\" (UniqueName: \"kubernetes.io/projected/1b324875-8f1a-41f5-b9e7-63c004cebba3-kube-api-access-9l8zl\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.707866 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b324875-8f1a-41f5-b9e7-63c004cebba3-logs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.712443 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-config-data\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.714209 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.718508 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-public-tls-certs\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.719890 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.726927 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.733302 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l8zl\" (UniqueName: \"kubernetes.io/projected/1b324875-8f1a-41f5-b9e7-63c004cebba3-kube-api-access-9l8zl\") pod \"nova-api-0\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " pod="openstack/nova-api-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.766210 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:32 crc kubenswrapper[4957]: I1206 06:05:32.774229 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.284042 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.373777 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b324875-8f1a-41f5-b9e7-63c004cebba3","Type":"ContainerStarted","Data":"343f4f73b093e322540c1d6dd00d3cf1d671d5a66049e3e42efb89bb386813fc"} Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.376589 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerStarted","Data":"adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83"} Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.395959 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.662566 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-56t98"] Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.664241 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.666321 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.666782 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.688992 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-56t98"] Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.712137 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.712411 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.829453 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-scripts\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.829506 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-config-data\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.829571 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cjzr\" (UniqueName: \"kubernetes.io/projected/9c5dbab8-a215-48b4-beea-7a5ae2715156-kube-api-access-8cjzr\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.829735 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.932533 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.932630 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-scripts\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.932654 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-config-data\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.932700 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cjzr\" (UniqueName: \"kubernetes.io/projected/9c5dbab8-a215-48b4-beea-7a5ae2715156-kube-api-access-8cjzr\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.938997 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-scripts\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.939362 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.943462 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-config-data\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:33 crc kubenswrapper[4957]: I1206 06:05:33.956234 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cjzr\" (UniqueName: \"kubernetes.io/projected/9c5dbab8-a215-48b4-beea-7a5ae2715156-kube-api-access-8cjzr\") pod \"nova-cell1-cell-mapping-56t98\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.037046 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.389614 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerStarted","Data":"f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d"} Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.394303 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b324875-8f1a-41f5-b9e7-63c004cebba3","Type":"ContainerStarted","Data":"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6"} Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.394361 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b324875-8f1a-41f5-b9e7-63c004cebba3","Type":"ContainerStarted","Data":"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660"} Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.496562 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.496539361 podStartE2EDuration="2.496539361s" podCreationTimestamp="2025-12-06 06:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:34.417546175 +0000 UTC m=+1574.067813827" watchObservedRunningTime="2025-12-06 06:05:34.496539361 +0000 UTC m=+1574.146807003" Dec 06 06:05:34 crc kubenswrapper[4957]: W1206 06:05:34.499183 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c5dbab8_a215_48b4_beea_7a5ae2715156.slice/crio-3219766967ae80b0791d4ea9dfc6bd60d00ca689f5d631fd38380cff17893dcf WatchSource:0}: Error finding container 3219766967ae80b0791d4ea9dfc6bd60d00ca689f5d631fd38380cff17893dcf: Status 404 returned error can't find the container with id 3219766967ae80b0791d4ea9dfc6bd60d00ca689f5d631fd38380cff17893dcf Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.506137 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-56t98"] Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.839996 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.915946 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-vx2p7"] Dec 06 06:05:34 crc kubenswrapper[4957]: I1206 06:05:34.916227 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" podUID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerName="dnsmasq-dns" containerID="cri-o://8867c86d2b2990d2807f29679c6133960ebd025317ef5b7906227a45aed90f06" gracePeriod=10 Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.406703 4957 generic.go:334] "Generic (PLEG): container finished" podID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerID="8867c86d2b2990d2807f29679c6133960ebd025317ef5b7906227a45aed90f06" exitCode=0 Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.406794 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" event={"ID":"296035c8-c16e-4a85-b712-78f8ee8f70cd","Type":"ContainerDied","Data":"8867c86d2b2990d2807f29679c6133960ebd025317ef5b7906227a45aed90f06"} Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.407046 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" event={"ID":"296035c8-c16e-4a85-b712-78f8ee8f70cd","Type":"ContainerDied","Data":"90e931ae681b26eeaf323294c2cd78dbb9c89c69bfff55f1842d9595350f33dc"} Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.407062 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90e931ae681b26eeaf323294c2cd78dbb9c89c69bfff55f1842d9595350f33dc" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.413866 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerStarted","Data":"1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476"} Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.413946 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-central-agent" containerID="cri-o://e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc" gracePeriod=30 Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.413988 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.414047 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="proxy-httpd" containerID="cri-o://1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476" gracePeriod=30 Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.414096 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="sg-core" containerID="cri-o://f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d" gracePeriod=30 Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.414135 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-notification-agent" containerID="cri-o://adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83" gracePeriod=30 Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.419547 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-56t98" event={"ID":"9c5dbab8-a215-48b4-beea-7a5ae2715156","Type":"ContainerStarted","Data":"e6b39df616f7b247d0534fa5dbdc7b2cc9c3de3a6b32187da0775b5c1e0a0c72"} Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.419591 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-56t98" event={"ID":"9c5dbab8-a215-48b4-beea-7a5ae2715156","Type":"ContainerStarted","Data":"3219766967ae80b0791d4ea9dfc6bd60d00ca689f5d631fd38380cff17893dcf"} Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.422240 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.451475 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.81529241 podStartE2EDuration="5.451451308s" podCreationTimestamp="2025-12-06 06:05:30 +0000 UTC" firstStartedPulling="2025-12-06 06:05:31.351389456 +0000 UTC m=+1571.001657088" lastFinishedPulling="2025-12-06 06:05:34.987548354 +0000 UTC m=+1574.637815986" observedRunningTime="2025-12-06 06:05:35.444113428 +0000 UTC m=+1575.094381080" watchObservedRunningTime="2025-12-06 06:05:35.451451308 +0000 UTC m=+1575.101718940" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.464910 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-56t98" podStartSLOduration=2.464889175 podStartE2EDuration="2.464889175s" podCreationTimestamp="2025-12-06 06:05:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:35.457314888 +0000 UTC m=+1575.107582530" watchObservedRunningTime="2025-12-06 06:05:35.464889175 +0000 UTC m=+1575.115156807" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.574393 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-sb\") pod \"296035c8-c16e-4a85-b712-78f8ee8f70cd\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.574480 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-nb\") pod \"296035c8-c16e-4a85-b712-78f8ee8f70cd\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.574591 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd2nk\" (UniqueName: \"kubernetes.io/projected/296035c8-c16e-4a85-b712-78f8ee8f70cd-kube-api-access-pd2nk\") pod \"296035c8-c16e-4a85-b712-78f8ee8f70cd\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.574690 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-svc\") pod \"296035c8-c16e-4a85-b712-78f8ee8f70cd\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.574742 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-config\") pod \"296035c8-c16e-4a85-b712-78f8ee8f70cd\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.574786 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-swift-storage-0\") pod \"296035c8-c16e-4a85-b712-78f8ee8f70cd\" (UID: \"296035c8-c16e-4a85-b712-78f8ee8f70cd\") " Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.582581 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296035c8-c16e-4a85-b712-78f8ee8f70cd-kube-api-access-pd2nk" (OuterVolumeSpecName: "kube-api-access-pd2nk") pod "296035c8-c16e-4a85-b712-78f8ee8f70cd" (UID: "296035c8-c16e-4a85-b712-78f8ee8f70cd"). InnerVolumeSpecName "kube-api-access-pd2nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.627900 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "296035c8-c16e-4a85-b712-78f8ee8f70cd" (UID: "296035c8-c16e-4a85-b712-78f8ee8f70cd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.629141 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "296035c8-c16e-4a85-b712-78f8ee8f70cd" (UID: "296035c8-c16e-4a85-b712-78f8ee8f70cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.629355 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "296035c8-c16e-4a85-b712-78f8ee8f70cd" (UID: "296035c8-c16e-4a85-b712-78f8ee8f70cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.636350 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "296035c8-c16e-4a85-b712-78f8ee8f70cd" (UID: "296035c8-c16e-4a85-b712-78f8ee8f70cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.642455 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-config" (OuterVolumeSpecName: "config") pod "296035c8-c16e-4a85-b712-78f8ee8f70cd" (UID: "296035c8-c16e-4a85-b712-78f8ee8f70cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.677472 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.677513 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.677527 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd2nk\" (UniqueName: \"kubernetes.io/projected/296035c8-c16e-4a85-b712-78f8ee8f70cd-kube-api-access-pd2nk\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.677539 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.677550 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:35 crc kubenswrapper[4957]: I1206 06:05:35.677562 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296035c8-c16e-4a85-b712-78f8ee8f70cd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.432008 4957 generic.go:334] "Generic (PLEG): container finished" podID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerID="f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d" exitCode=2 Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.432047 4957 generic.go:334] "Generic (PLEG): container finished" podID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerID="adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83" exitCode=0 Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.432140 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerDied","Data":"f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d"} Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.432199 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerDied","Data":"adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83"} Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.433368 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-vx2p7" Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.481650 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-vx2p7"] Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.490061 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-vx2p7"] Dec 06 06:05:36 crc kubenswrapper[4957]: I1206 06:05:36.673757 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="296035c8-c16e-4a85-b712-78f8ee8f70cd" path="/var/lib/kubelet/pods/296035c8-c16e-4a85-b712-78f8ee8f70cd/volumes" Dec 06 06:05:39 crc kubenswrapper[4957]: I1206 06:05:39.488308 4957 generic.go:334] "Generic (PLEG): container finished" podID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerID="e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc" exitCode=0 Dec 06 06:05:39 crc kubenswrapper[4957]: I1206 06:05:39.488413 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerDied","Data":"e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc"} Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.181451 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.181806 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.182031 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.183011 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.183184 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" gracePeriod=600 Dec 06 06:05:40 crc kubenswrapper[4957]: E1206 06:05:40.307257 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.499352 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" exitCode=0 Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.499414 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc"} Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.499452 4957 scope.go:117] "RemoveContainer" containerID="b3b7a79d60429782e416177d4e1df59e205a0e6bbe183ad64871d22be42eb374" Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.500006 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:05:40 crc kubenswrapper[4957]: E1206 06:05:40.500338 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.501804 4957 generic.go:334] "Generic (PLEG): container finished" podID="9c5dbab8-a215-48b4-beea-7a5ae2715156" containerID="e6b39df616f7b247d0534fa5dbdc7b2cc9c3de3a6b32187da0775b5c1e0a0c72" exitCode=0 Dec 06 06:05:40 crc kubenswrapper[4957]: I1206 06:05:40.501841 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-56t98" event={"ID":"9c5dbab8-a215-48b4-beea-7a5ae2715156","Type":"ContainerDied","Data":"e6b39df616f7b247d0534fa5dbdc7b2cc9c3de3a6b32187da0775b5c1e0a0c72"} Dec 06 06:05:41 crc kubenswrapper[4957]: I1206 06:05:41.855061 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.008411 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-scripts\") pod \"9c5dbab8-a215-48b4-beea-7a5ae2715156\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.008462 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-combined-ca-bundle\") pod \"9c5dbab8-a215-48b4-beea-7a5ae2715156\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.008527 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-config-data\") pod \"9c5dbab8-a215-48b4-beea-7a5ae2715156\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.008677 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cjzr\" (UniqueName: \"kubernetes.io/projected/9c5dbab8-a215-48b4-beea-7a5ae2715156-kube-api-access-8cjzr\") pod \"9c5dbab8-a215-48b4-beea-7a5ae2715156\" (UID: \"9c5dbab8-a215-48b4-beea-7a5ae2715156\") " Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.013783 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-scripts" (OuterVolumeSpecName: "scripts") pod "9c5dbab8-a215-48b4-beea-7a5ae2715156" (UID: "9c5dbab8-a215-48b4-beea-7a5ae2715156"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.014216 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5dbab8-a215-48b4-beea-7a5ae2715156-kube-api-access-8cjzr" (OuterVolumeSpecName: "kube-api-access-8cjzr") pod "9c5dbab8-a215-48b4-beea-7a5ae2715156" (UID: "9c5dbab8-a215-48b4-beea-7a5ae2715156"). InnerVolumeSpecName "kube-api-access-8cjzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.037200 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-config-data" (OuterVolumeSpecName: "config-data") pod "9c5dbab8-a215-48b4-beea-7a5ae2715156" (UID: "9c5dbab8-a215-48b4-beea-7a5ae2715156"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.046782 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c5dbab8-a215-48b4-beea-7a5ae2715156" (UID: "9c5dbab8-a215-48b4-beea-7a5ae2715156"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.110910 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.110962 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.110978 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c5dbab8-a215-48b4-beea-7a5ae2715156-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.110992 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cjzr\" (UniqueName: \"kubernetes.io/projected/9c5dbab8-a215-48b4-beea-7a5ae2715156-kube-api-access-8cjzr\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.523905 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-56t98" event={"ID":"9c5dbab8-a215-48b4-beea-7a5ae2715156","Type":"ContainerDied","Data":"3219766967ae80b0791d4ea9dfc6bd60d00ca689f5d631fd38380cff17893dcf"} Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.523944 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3219766967ae80b0791d4ea9dfc6bd60d00ca689f5d631fd38380cff17893dcf" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.523974 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-56t98" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.709899 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.710150 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-log" containerID="cri-o://47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660" gracePeriod=30 Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.710230 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-api" containerID="cri-o://f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6" gracePeriod=30 Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.721908 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.722461 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.722636 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b3b61136-8d49-4883-b52c-f370b552d208" containerName="nova-scheduler-scheduler" containerID="cri-o://e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5" gracePeriod=30 Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.762532 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.771575 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:42 crc kubenswrapper[4957]: I1206 06:05:42.780819 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.413039 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.436589 4957 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.438332 4957 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.444391 4957 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.444452 4957 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b3b61136-8d49-4883-b52c-f370b552d208" containerName="nova-scheduler-scheduler" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.534332 4957 generic.go:334] "Generic (PLEG): container finished" podID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerID="f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6" exitCode=0 Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.534362 4957 generic.go:334] "Generic (PLEG): container finished" podID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerID="47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660" exitCode=143 Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.534378 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b324875-8f1a-41f5-b9e7-63c004cebba3","Type":"ContainerDied","Data":"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6"} Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.534416 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.534435 4957 scope.go:117] "RemoveContainer" containerID="f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.534424 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b324875-8f1a-41f5-b9e7-63c004cebba3","Type":"ContainerDied","Data":"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660"} Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.534562 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b324875-8f1a-41f5-b9e7-63c004cebba3","Type":"ContainerDied","Data":"343f4f73b093e322540c1d6dd00d3cf1d671d5a66049e3e42efb89bb386813fc"} Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.536116 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l8zl\" (UniqueName: \"kubernetes.io/projected/1b324875-8f1a-41f5-b9e7-63c004cebba3-kube-api-access-9l8zl\") pod \"1b324875-8f1a-41f5-b9e7-63c004cebba3\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.536235 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b324875-8f1a-41f5-b9e7-63c004cebba3-logs\") pod \"1b324875-8f1a-41f5-b9e7-63c004cebba3\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.536297 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-config-data\") pod \"1b324875-8f1a-41f5-b9e7-63c004cebba3\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.536327 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-public-tls-certs\") pod \"1b324875-8f1a-41f5-b9e7-63c004cebba3\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.536382 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-combined-ca-bundle\") pod \"1b324875-8f1a-41f5-b9e7-63c004cebba3\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.536419 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-internal-tls-certs\") pod \"1b324875-8f1a-41f5-b9e7-63c004cebba3\" (UID: \"1b324875-8f1a-41f5-b9e7-63c004cebba3\") " Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.537642 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b324875-8f1a-41f5-b9e7-63c004cebba3-logs" (OuterVolumeSpecName: "logs") pod "1b324875-8f1a-41f5-b9e7-63c004cebba3" (UID: "1b324875-8f1a-41f5-b9e7-63c004cebba3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.544013 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b324875-8f1a-41f5-b9e7-63c004cebba3-kube-api-access-9l8zl" (OuterVolumeSpecName: "kube-api-access-9l8zl") pod "1b324875-8f1a-41f5-b9e7-63c004cebba3" (UID: "1b324875-8f1a-41f5-b9e7-63c004cebba3"). InnerVolumeSpecName "kube-api-access-9l8zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.550632 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.575363 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b324875-8f1a-41f5-b9e7-63c004cebba3" (UID: "1b324875-8f1a-41f5-b9e7-63c004cebba3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.617729 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-config-data" (OuterVolumeSpecName: "config-data") pod "1b324875-8f1a-41f5-b9e7-63c004cebba3" (UID: "1b324875-8f1a-41f5-b9e7-63c004cebba3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.650134 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1b324875-8f1a-41f5-b9e7-63c004cebba3" (UID: "1b324875-8f1a-41f5-b9e7-63c004cebba3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.651515 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.651552 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.651566 4957 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.651580 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l8zl\" (UniqueName: \"kubernetes.io/projected/1b324875-8f1a-41f5-b9e7-63c004cebba3-kube-api-access-9l8zl\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.651594 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b324875-8f1a-41f5-b9e7-63c004cebba3-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.654062 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1b324875-8f1a-41f5-b9e7-63c004cebba3" (UID: "1b324875-8f1a-41f5-b9e7-63c004cebba3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.668670 4957 scope.go:117] "RemoveContainer" containerID="47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.696773 4957 scope.go:117] "RemoveContainer" containerID="f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6" Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.703124 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6\": container with ID starting with f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6 not found: ID does not exist" containerID="f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.703276 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6"} err="failed to get container status \"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6\": rpc error: code = NotFound desc = could not find container \"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6\": container with ID starting with f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6 not found: ID does not exist" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.703393 4957 scope.go:117] "RemoveContainer" containerID="47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660" Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.703753 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660\": container with ID starting with 47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660 not found: ID does not exist" containerID="47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.703878 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660"} err="failed to get container status \"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660\": rpc error: code = NotFound desc = could not find container \"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660\": container with ID starting with 47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660 not found: ID does not exist" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.703968 4957 scope.go:117] "RemoveContainer" containerID="f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.704742 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6"} err="failed to get container status \"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6\": rpc error: code = NotFound desc = could not find container \"f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6\": container with ID starting with f7e43e8eac7679f493b776978fee56e7c13303bdc7df412a7b10a9716ebb19c6 not found: ID does not exist" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.704912 4957 scope.go:117] "RemoveContainer" containerID="47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.708982 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660"} err="failed to get container status \"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660\": rpc error: code = NotFound desc = could not find container \"47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660\": container with ID starting with 47b4c992e611da1b60d3808ae4acb6236070542cd3f2492e2c3fd4b56dd57660 not found: ID does not exist" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.753368 4957 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b324875-8f1a-41f5-b9e7-63c004cebba3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.938230 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.957157 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970162 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.970557 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerName="dnsmasq-dns" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970574 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerName="dnsmasq-dns" Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.970596 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c5dbab8-a215-48b4-beea-7a5ae2715156" containerName="nova-manage" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970602 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c5dbab8-a215-48b4-beea-7a5ae2715156" containerName="nova-manage" Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.970611 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-log" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970618 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-log" Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.970629 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-api" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970634 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-api" Dec 06 06:05:43 crc kubenswrapper[4957]: E1206 06:05:43.970643 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerName="init" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970648 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerName="init" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970842 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-api" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970857 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" containerName="nova-api-log" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970875 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="296035c8-c16e-4a85-b712-78f8ee8f70cd" containerName="dnsmasq-dns" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.970882 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c5dbab8-a215-48b4-beea-7a5ae2715156" containerName="nova-manage" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.972588 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.976385 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.976650 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.977101 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 06:05:43 crc kubenswrapper[4957]: I1206 06:05:43.985061 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.062093 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.062224 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-config-data\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.062297 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.062336 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ae0f500-0ce2-4374-afe8-cc65e465ac26-logs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.062362 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-public-tls-certs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.062418 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtrqr\" (UniqueName: \"kubernetes.io/projected/1ae0f500-0ce2-4374-afe8-cc65e465ac26-kube-api-access-wtrqr\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.163776 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-config-data\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.164129 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.164232 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ae0f500-0ce2-4374-afe8-cc65e465ac26-logs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.164364 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-public-tls-certs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.165097 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtrqr\" (UniqueName: \"kubernetes.io/projected/1ae0f500-0ce2-4374-afe8-cc65e465ac26-kube-api-access-wtrqr\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.165252 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.164661 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ae0f500-0ce2-4374-afe8-cc65e465ac26-logs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.168267 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-public-tls-certs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.168279 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-config-data\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.168720 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.169915 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ae0f500-0ce2-4374-afe8-cc65e465ac26-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.182968 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtrqr\" (UniqueName: \"kubernetes.io/projected/1ae0f500-0ce2-4374-afe8-cc65e465ac26-kube-api-access-wtrqr\") pod \"nova-api-0\" (UID: \"1ae0f500-0ce2-4374-afe8-cc65e465ac26\") " pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.300754 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.547426 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-log" containerID="cri-o://196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9" gracePeriod=30 Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.547707 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-metadata" containerID="cri-o://183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36" gracePeriod=30 Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.678539 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b324875-8f1a-41f5-b9e7-63c004cebba3" path="/var/lib/kubelet/pods/1b324875-8f1a-41f5-b9e7-63c004cebba3/volumes" Dec 06 06:05:44 crc kubenswrapper[4957]: I1206 06:05:44.848159 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 06:05:44 crc kubenswrapper[4957]: W1206 06:05:44.864775 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ae0f500_0ce2_4374_afe8_cc65e465ac26.slice/crio-cbf504fe1b06d09e5b7c047f56d4210c1d01eb5f08eada6dbcfd4c434171c6de WatchSource:0}: Error finding container cbf504fe1b06d09e5b7c047f56d4210c1d01eb5f08eada6dbcfd4c434171c6de: Status 404 returned error can't find the container with id cbf504fe1b06d09e5b7c047f56d4210c1d01eb5f08eada6dbcfd4c434171c6de Dec 06 06:05:45 crc kubenswrapper[4957]: I1206 06:05:45.559459 4957 generic.go:334] "Generic (PLEG): container finished" podID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerID="196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9" exitCode=143 Dec 06 06:05:45 crc kubenswrapper[4957]: I1206 06:05:45.559557 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af7dc285-d8d8-4d28-9335-9f0abaa31738","Type":"ContainerDied","Data":"196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9"} Dec 06 06:05:45 crc kubenswrapper[4957]: I1206 06:05:45.562579 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ae0f500-0ce2-4374-afe8-cc65e465ac26","Type":"ContainerStarted","Data":"b9882bc21c7b6afb12afcac877b4053af6148b2a49c66fe06786a9d492fe4645"} Dec 06 06:05:45 crc kubenswrapper[4957]: I1206 06:05:45.562685 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ae0f500-0ce2-4374-afe8-cc65e465ac26","Type":"ContainerStarted","Data":"cbf504fe1b06d09e5b7c047f56d4210c1d01eb5f08eada6dbcfd4c434171c6de"} Dec 06 06:05:46 crc kubenswrapper[4957]: I1206 06:05:46.574711 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ae0f500-0ce2-4374-afe8-cc65e465ac26","Type":"ContainerStarted","Data":"b1846fc106a6a06219e70fc31b52ffede034f243fa6ff3774735091d522671c6"} Dec 06 06:05:46 crc kubenswrapper[4957]: I1206 06:05:46.607704 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.607682656 podStartE2EDuration="3.607682656s" podCreationTimestamp="2025-12-06 06:05:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:46.600254644 +0000 UTC m=+1586.250522296" watchObservedRunningTime="2025-12-06 06:05:46.607682656 +0000 UTC m=+1586.257950288" Dec 06 06:05:47 crc kubenswrapper[4957]: I1206 06:05:47.585677 4957 generic.go:334] "Generic (PLEG): container finished" podID="b3b61136-8d49-4883-b52c-f370b552d208" containerID="e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5" exitCode=0 Dec 06 06:05:47 crc kubenswrapper[4957]: I1206 06:05:47.586796 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3b61136-8d49-4883-b52c-f370b552d208","Type":"ContainerDied","Data":"e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5"} Dec 06 06:05:47 crc kubenswrapper[4957]: I1206 06:05:47.697128 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": dial tcp 10.217.0.197:8775: connect: connection refused" Dec 06 06:05:47 crc kubenswrapper[4957]: I1206 06:05:47.697347 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": dial tcp 10.217.0.197:8775: connect: connection refused" Dec 06 06:05:47 crc kubenswrapper[4957]: I1206 06:05:47.914187 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.014334 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.047534 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-config-data\") pod \"b3b61136-8d49-4883-b52c-f370b552d208\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.047667 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rs2h\" (UniqueName: \"kubernetes.io/projected/b3b61136-8d49-4883-b52c-f370b552d208-kube-api-access-5rs2h\") pod \"b3b61136-8d49-4883-b52c-f370b552d208\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.047717 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-combined-ca-bundle\") pod \"b3b61136-8d49-4883-b52c-f370b552d208\" (UID: \"b3b61136-8d49-4883-b52c-f370b552d208\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.074351 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3b61136-8d49-4883-b52c-f370b552d208-kube-api-access-5rs2h" (OuterVolumeSpecName: "kube-api-access-5rs2h") pod "b3b61136-8d49-4883-b52c-f370b552d208" (UID: "b3b61136-8d49-4883-b52c-f370b552d208"). InnerVolumeSpecName "kube-api-access-5rs2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.087471 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-config-data" (OuterVolumeSpecName: "config-data") pod "b3b61136-8d49-4883-b52c-f370b552d208" (UID: "b3b61136-8d49-4883-b52c-f370b552d208"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.113143 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3b61136-8d49-4883-b52c-f370b552d208" (UID: "b3b61136-8d49-4883-b52c-f370b552d208"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.149400 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-config-data\") pod \"af7dc285-d8d8-4d28-9335-9f0abaa31738\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.149522 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-combined-ca-bundle\") pod \"af7dc285-d8d8-4d28-9335-9f0abaa31738\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.149684 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7dc285-d8d8-4d28-9335-9f0abaa31738-logs\") pod \"af7dc285-d8d8-4d28-9335-9f0abaa31738\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.149717 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkgl8\" (UniqueName: \"kubernetes.io/projected/af7dc285-d8d8-4d28-9335-9f0abaa31738-kube-api-access-fkgl8\") pod \"af7dc285-d8d8-4d28-9335-9f0abaa31738\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.149758 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-nova-metadata-tls-certs\") pod \"af7dc285-d8d8-4d28-9335-9f0abaa31738\" (UID: \"af7dc285-d8d8-4d28-9335-9f0abaa31738\") " Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.150219 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.150238 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rs2h\" (UniqueName: \"kubernetes.io/projected/b3b61136-8d49-4883-b52c-f370b552d208-kube-api-access-5rs2h\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.150248 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b61136-8d49-4883-b52c-f370b552d208-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.151149 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af7dc285-d8d8-4d28-9335-9f0abaa31738-logs" (OuterVolumeSpecName: "logs") pod "af7dc285-d8d8-4d28-9335-9f0abaa31738" (UID: "af7dc285-d8d8-4d28-9335-9f0abaa31738"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.153430 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af7dc285-d8d8-4d28-9335-9f0abaa31738-kube-api-access-fkgl8" (OuterVolumeSpecName: "kube-api-access-fkgl8") pod "af7dc285-d8d8-4d28-9335-9f0abaa31738" (UID: "af7dc285-d8d8-4d28-9335-9f0abaa31738"). InnerVolumeSpecName "kube-api-access-fkgl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.177906 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-config-data" (OuterVolumeSpecName: "config-data") pod "af7dc285-d8d8-4d28-9335-9f0abaa31738" (UID: "af7dc285-d8d8-4d28-9335-9f0abaa31738"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.179915 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af7dc285-d8d8-4d28-9335-9f0abaa31738" (UID: "af7dc285-d8d8-4d28-9335-9f0abaa31738"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.200500 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "af7dc285-d8d8-4d28-9335-9f0abaa31738" (UID: "af7dc285-d8d8-4d28-9335-9f0abaa31738"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.252104 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.252145 4957 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af7dc285-d8d8-4d28-9335-9f0abaa31738-logs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.252158 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkgl8\" (UniqueName: \"kubernetes.io/projected/af7dc285-d8d8-4d28-9335-9f0abaa31738-kube-api-access-fkgl8\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.252168 4957 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.252177 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af7dc285-d8d8-4d28-9335-9f0abaa31738-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.603221 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.603212 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3b61136-8d49-4883-b52c-f370b552d208","Type":"ContainerDied","Data":"11606f6c949cdc6f9c52fdb27c32406f82b550638b6dd55bc1380a5e5b4ae524"} Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.603409 4957 scope.go:117] "RemoveContainer" containerID="e4f392058806ed48fe58e45a8a3ff4fbd5c1e40a3c2bd935f946a2dcbf48b4f5" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.606922 4957 generic.go:334] "Generic (PLEG): container finished" podID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerID="183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36" exitCode=0 Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.606958 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af7dc285-d8d8-4d28-9335-9f0abaa31738","Type":"ContainerDied","Data":"183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36"} Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.606986 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af7dc285-d8d8-4d28-9335-9f0abaa31738","Type":"ContainerDied","Data":"329cdff5540f679ae3a7368e72c4b9c1965f07a122d6d800c2a57a268f9711a7"} Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.607042 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.631721 4957 scope.go:117] "RemoveContainer" containerID="183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.649284 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.660776 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.677222 4957 scope.go:117] "RemoveContainer" containerID="196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.689076 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3b61136-8d49-4883-b52c-f370b552d208" path="/var/lib/kubelet/pods/b3b61136-8d49-4883-b52c-f370b552d208/volumes" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.689735 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.704824 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.713801 4957 scope.go:117] "RemoveContainer" containerID="183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36" Dec 06 06:05:48 crc kubenswrapper[4957]: E1206 06:05:48.714767 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36\": container with ID starting with 183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36 not found: ID does not exist" containerID="183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.714814 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36"} err="failed to get container status \"183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36\": rpc error: code = NotFound desc = could not find container \"183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36\": container with ID starting with 183dec88c5023fc9deb545a8eee15b61956c250e7cdf91b595617f92212b4a36 not found: ID does not exist" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.714857 4957 scope.go:117] "RemoveContainer" containerID="196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9" Dec 06 06:05:48 crc kubenswrapper[4957]: E1206 06:05:48.715105 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9\": container with ID starting with 196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9 not found: ID does not exist" containerID="196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.715130 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9"} err="failed to get container status \"196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9\": rpc error: code = NotFound desc = could not find container \"196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9\": container with ID starting with 196b351eb5b32f60c38a2361db2cac97291db3f753d5b8e90f9fe0a36877f2b9 not found: ID does not exist" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.720055 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: E1206 06:05:48.720450 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-log" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.720463 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-log" Dec 06 06:05:48 crc kubenswrapper[4957]: E1206 06:05:48.720484 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b61136-8d49-4883-b52c-f370b552d208" containerName="nova-scheduler-scheduler" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.720491 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b61136-8d49-4883-b52c-f370b552d208" containerName="nova-scheduler-scheduler" Dec 06 06:05:48 crc kubenswrapper[4957]: E1206 06:05:48.720502 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-metadata" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.720508 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-metadata" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.720670 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-metadata" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.720687 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" containerName="nova-metadata-log" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.720699 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b61136-8d49-4883-b52c-f370b552d208" containerName="nova-scheduler-scheduler" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.721582 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.724181 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.724573 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.737366 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.741131 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.743424 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.757802 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.785306 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.867598 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-logs\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.867661 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-config-data\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.867747 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c683a5a8-170a-4880-b9fa-7b9635ce0751-config-data\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.867796 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c683a5a8-170a-4880-b9fa-7b9635ce0751-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.867906 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c49jd\" (UniqueName: \"kubernetes.io/projected/c683a5a8-170a-4880-b9fa-7b9635ce0751-kube-api-access-c49jd\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.867949 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzb85\" (UniqueName: \"kubernetes.io/projected/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-kube-api-access-qzb85\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.867993 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.868065 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969360 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969418 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969502 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-logs\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969521 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-config-data\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969542 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c683a5a8-170a-4880-b9fa-7b9635ce0751-config-data\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969576 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c683a5a8-170a-4880-b9fa-7b9635ce0751-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969612 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c49jd\" (UniqueName: \"kubernetes.io/projected/c683a5a8-170a-4880-b9fa-7b9635ce0751-kube-api-access-c49jd\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.969636 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzb85\" (UniqueName: \"kubernetes.io/projected/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-kube-api-access-qzb85\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.970114 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-logs\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.975402 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-config-data\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.975522 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c683a5a8-170a-4880-b9fa-7b9635ce0751-config-data\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.977495 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.985272 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.988424 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c683a5a8-170a-4880-b9fa-7b9635ce0751-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.989600 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c49jd\" (UniqueName: \"kubernetes.io/projected/c683a5a8-170a-4880-b9fa-7b9635ce0751-kube-api-access-c49jd\") pod \"nova-scheduler-0\" (UID: \"c683a5a8-170a-4880-b9fa-7b9635ce0751\") " pod="openstack/nova-scheduler-0" Dec 06 06:05:48 crc kubenswrapper[4957]: I1206 06:05:48.990046 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzb85\" (UniqueName: \"kubernetes.io/projected/67d6edc5-c0f1-4466-8b5b-99cd3bc06d17-kube-api-access-qzb85\") pod \"nova-metadata-0\" (UID: \"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17\") " pod="openstack/nova-metadata-0" Dec 06 06:05:49 crc kubenswrapper[4957]: I1206 06:05:49.044965 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 06:05:49 crc kubenswrapper[4957]: I1206 06:05:49.084949 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 06:05:49 crc kubenswrapper[4957]: I1206 06:05:49.609159 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 06:05:49 crc kubenswrapper[4957]: I1206 06:05:49.626252 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c683a5a8-170a-4880-b9fa-7b9635ce0751","Type":"ContainerStarted","Data":"a3427b53ca584999b05fe1b5909e87630738240029c76dab1e155b41b3240d78"} Dec 06 06:05:49 crc kubenswrapper[4957]: I1206 06:05:49.628817 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 06:05:50 crc kubenswrapper[4957]: I1206 06:05:50.657141 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c683a5a8-170a-4880-b9fa-7b9635ce0751","Type":"ContainerStarted","Data":"2c74838e3d58df640d47742a825050b11e577089632fed50e30e121703546b35"} Dec 06 06:05:50 crc kubenswrapper[4957]: I1206 06:05:50.691078 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af7dc285-d8d8-4d28-9335-9f0abaa31738" path="/var/lib/kubelet/pods/af7dc285-d8d8-4d28-9335-9f0abaa31738/volumes" Dec 06 06:05:50 crc kubenswrapper[4957]: I1206 06:05:50.695577 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17","Type":"ContainerStarted","Data":"fa6838a15bd66739bcf7619d73f10bcdf78645e723c15009775ede67b9c97f39"} Dec 06 06:05:50 crc kubenswrapper[4957]: I1206 06:05:50.696006 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17","Type":"ContainerStarted","Data":"085ade475bd30a75075d38e89907634948dfac05de7e1b25959b7a0e9b1beddb"} Dec 06 06:05:51 crc kubenswrapper[4957]: I1206 06:05:51.673582 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"67d6edc5-c0f1-4466-8b5b-99cd3bc06d17","Type":"ContainerStarted","Data":"995c53cfdef58d0311316269d4574c16035821bc0983ca0e9f6b1be6944ababd"} Dec 06 06:05:51 crc kubenswrapper[4957]: I1206 06:05:51.691558 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.691540814 podStartE2EDuration="3.691540814s" podCreationTimestamp="2025-12-06 06:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:51.687939346 +0000 UTC m=+1591.338206978" watchObservedRunningTime="2025-12-06 06:05:51.691540814 +0000 UTC m=+1591.341808456" Dec 06 06:05:54 crc kubenswrapper[4957]: I1206 06:05:54.045113 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:05:54 crc kubenswrapper[4957]: I1206 06:05:54.045446 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 06:05:54 crc kubenswrapper[4957]: I1206 06:05:54.085629 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 06:05:54 crc kubenswrapper[4957]: I1206 06:05:54.301911 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:05:54 crc kubenswrapper[4957]: I1206 06:05:54.301986 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 06:05:54 crc kubenswrapper[4957]: I1206 06:05:54.662552 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:05:54 crc kubenswrapper[4957]: E1206 06:05:54.662898 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:05:55 crc kubenswrapper[4957]: I1206 06:05:55.315003 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1ae0f500-0ce2-4374-afe8-cc65e465ac26" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:05:55 crc kubenswrapper[4957]: I1206 06:05:55.315018 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1ae0f500-0ce2-4374-afe8-cc65e465ac26" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:05:59 crc kubenswrapper[4957]: I1206 06:05:59.045117 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:05:59 crc kubenswrapper[4957]: I1206 06:05:59.045488 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 06:05:59 crc kubenswrapper[4957]: I1206 06:05:59.086081 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 06:05:59 crc kubenswrapper[4957]: I1206 06:05:59.115613 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 06:05:59 crc kubenswrapper[4957]: I1206 06:05:59.141880 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=11.14186295 podStartE2EDuration="11.14186295s" podCreationTimestamp="2025-12-06 06:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:05:51.703320965 +0000 UTC m=+1591.353588607" watchObservedRunningTime="2025-12-06 06:05:59.14186295 +0000 UTC m=+1598.792130582" Dec 06 06:05:59 crc kubenswrapper[4957]: I1206 06:05:59.842864 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 06:06:00 crc kubenswrapper[4957]: I1206 06:06:00.056961 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="67d6edc5-c0f1-4466-8b5b-99cd3bc06d17" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:06:00 crc kubenswrapper[4957]: I1206 06:06:00.057617 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="67d6edc5-c0f1-4466-8b5b-99cd3bc06d17" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 06:06:00 crc kubenswrapper[4957]: I1206 06:06:00.779876 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 06 06:06:04 crc kubenswrapper[4957]: I1206 06:06:04.311643 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:06:04 crc kubenswrapper[4957]: I1206 06:06:04.312948 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:06:04 crc kubenswrapper[4957]: I1206 06:06:04.319867 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 06:06:04 crc kubenswrapper[4957]: I1206 06:06:04.322612 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:06:04 crc kubenswrapper[4957]: I1206 06:06:04.854853 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 06:06:04 crc kubenswrapper[4957]: I1206 06:06:04.864307 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.833538 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.877366 4957 generic.go:334] "Generic (PLEG): container finished" podID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerID="1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476" exitCode=137 Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.877448 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.877495 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerDied","Data":"1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476"} Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.877528 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30a9c22f-d5a5-4794-9f6b-70a16656726a","Type":"ContainerDied","Data":"f6384d890b53464cca00650f32f0b2e1d51800a4e4593e074f1b67cec4ad15aa"} Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.877549 4957 scope.go:117] "RemoveContainer" containerID="1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.895405 4957 scope.go:117] "RemoveContainer" containerID="f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.914853 4957 scope.go:117] "RemoveContainer" containerID="adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.935955 4957 scope.go:117] "RemoveContainer" containerID="e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936013 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-scripts\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936163 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-combined-ca-bundle\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936239 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-config-data\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936342 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-run-httpd\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936375 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-log-httpd\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936431 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-ceilometer-tls-certs\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936457 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-sg-core-conf-yaml\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936497 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ww6c\" (UniqueName: \"kubernetes.io/projected/30a9c22f-d5a5-4794-9f6b-70a16656726a-kube-api-access-9ww6c\") pod \"30a9c22f-d5a5-4794-9f6b-70a16656726a\" (UID: \"30a9c22f-d5a5-4794-9f6b-70a16656726a\") " Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.936851 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.937887 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.938260 4957 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.938276 4957 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30a9c22f-d5a5-4794-9f6b-70a16656726a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.941566 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a9c22f-d5a5-4794-9f6b-70a16656726a-kube-api-access-9ww6c" (OuterVolumeSpecName: "kube-api-access-9ww6c") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "kube-api-access-9ww6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.941648 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-scripts" (OuterVolumeSpecName: "scripts") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.969047 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:06:05 crc kubenswrapper[4957]: I1206 06:06:05.999027 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.019881 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.037913 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-config-data" (OuterVolumeSpecName: "config-data") pod "30a9c22f-d5a5-4794-9f6b-70a16656726a" (UID: "30a9c22f-d5a5-4794-9f6b-70a16656726a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.045426 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.045467 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.045478 4957 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.045489 4957 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.045504 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ww6c\" (UniqueName: \"kubernetes.io/projected/30a9c22f-d5a5-4794-9f6b-70a16656726a-kube-api-access-9ww6c\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.045515 4957 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a9c22f-d5a5-4794-9f6b-70a16656726a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.066005 4957 scope.go:117] "RemoveContainer" containerID="1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476" Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.066761 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476\": container with ID starting with 1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476 not found: ID does not exist" containerID="1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.066812 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476"} err="failed to get container status \"1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476\": rpc error: code = NotFound desc = could not find container \"1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476\": container with ID starting with 1f8bcfa0c261ffab17a99b30be2a58dc828de4b9110eb42e9e807ecd5e0e9476 not found: ID does not exist" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.066886 4957 scope.go:117] "RemoveContainer" containerID="f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d" Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.067912 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d\": container with ID starting with f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d not found: ID does not exist" containerID="f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.068090 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d"} err="failed to get container status \"f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d\": rpc error: code = NotFound desc = could not find container \"f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d\": container with ID starting with f4f12f34224f4b80f7d515c6c13ea741eccc557fd9d3804f939640316069c40d not found: ID does not exist" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.068138 4957 scope.go:117] "RemoveContainer" containerID="adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83" Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.068517 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83\": container with ID starting with adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83 not found: ID does not exist" containerID="adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.068557 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83"} err="failed to get container status \"adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83\": rpc error: code = NotFound desc = could not find container \"adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83\": container with ID starting with adaf2b0f7b41e8b2af6444906efc98e29a50ae02afe2230dbf36eadb4397fb83 not found: ID does not exist" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.068575 4957 scope.go:117] "RemoveContainer" containerID="e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc" Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.068921 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc\": container with ID starting with e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc not found: ID does not exist" containerID="e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.068953 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc"} err="failed to get container status \"e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc\": rpc error: code = NotFound desc = could not find container \"e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc\": container with ID starting with e7d47af5f0c25cb09f290341e30f214401f4a91829e0c107651ade96682d10bc not found: ID does not exist" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.228407 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.239305 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.257036 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.257652 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="proxy-httpd" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.257682 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="proxy-httpd" Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.257718 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="sg-core" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.257729 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="sg-core" Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.257753 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-notification-agent" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.257766 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-notification-agent" Dec 06 06:06:06 crc kubenswrapper[4957]: E1206 06:06:06.257791 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-central-agent" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.257803 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-central-agent" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.258169 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="proxy-httpd" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.258200 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-notification-agent" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.258239 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="ceilometer-central-agent" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.258255 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" containerName="sg-core" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.260907 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.264872 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.265165 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.265364 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.289423 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350131 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/963f3a5a-72a1-4fed-9ca8-edc041e612c3-log-httpd\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350186 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350301 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350343 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/963f3a5a-72a1-4fed-9ca8-edc041e612c3-run-httpd\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350371 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjl2h\" (UniqueName: \"kubernetes.io/projected/963f3a5a-72a1-4fed-9ca8-edc041e612c3-kube-api-access-tjl2h\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350413 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-scripts\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350460 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.350489 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-config-data\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452421 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-scripts\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452514 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452552 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-config-data\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452643 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/963f3a5a-72a1-4fed-9ca8-edc041e612c3-log-httpd\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452670 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452730 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452807 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/963f3a5a-72a1-4fed-9ca8-edc041e612c3-run-httpd\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.452874 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjl2h\" (UniqueName: \"kubernetes.io/projected/963f3a5a-72a1-4fed-9ca8-edc041e612c3-kube-api-access-tjl2h\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.453650 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/963f3a5a-72a1-4fed-9ca8-edc041e612c3-run-httpd\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.453684 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/963f3a5a-72a1-4fed-9ca8-edc041e612c3-log-httpd\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.457762 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.457825 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.458934 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.461325 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-scripts\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.464414 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/963f3a5a-72a1-4fed-9ca8-edc041e612c3-config-data\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.470950 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjl2h\" (UniqueName: \"kubernetes.io/projected/963f3a5a-72a1-4fed-9ca8-edc041e612c3-kube-api-access-tjl2h\") pod \"ceilometer-0\" (UID: \"963f3a5a-72a1-4fed-9ca8-edc041e612c3\") " pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.583571 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 06:06:06 crc kubenswrapper[4957]: I1206 06:06:06.689392 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a9c22f-d5a5-4794-9f6b-70a16656726a" path="/var/lib/kubelet/pods/30a9c22f-d5a5-4794-9f6b-70a16656726a/volumes" Dec 06 06:06:07 crc kubenswrapper[4957]: I1206 06:06:07.062272 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 06:06:07 crc kubenswrapper[4957]: W1206 06:06:07.062383 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod963f3a5a_72a1_4fed_9ca8_edc041e612c3.slice/crio-333cb4dc52c102fa15f40befffa007f9a9493934c195a62d6b7b9f45feb3b2ce WatchSource:0}: Error finding container 333cb4dc52c102fa15f40befffa007f9a9493934c195a62d6b7b9f45feb3b2ce: Status 404 returned error can't find the container with id 333cb4dc52c102fa15f40befffa007f9a9493934c195a62d6b7b9f45feb3b2ce Dec 06 06:06:07 crc kubenswrapper[4957]: I1206 06:06:07.898181 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"963f3a5a-72a1-4fed-9ca8-edc041e612c3","Type":"ContainerStarted","Data":"0cc74ce9797c98b6f243621f7c207e7af951b6827b4851922186892b14895303"} Dec 06 06:06:07 crc kubenswrapper[4957]: I1206 06:06:07.898527 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"963f3a5a-72a1-4fed-9ca8-edc041e612c3","Type":"ContainerStarted","Data":"333cb4dc52c102fa15f40befffa007f9a9493934c195a62d6b7b9f45feb3b2ce"} Dec 06 06:06:08 crc kubenswrapper[4957]: I1206 06:06:08.910157 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"963f3a5a-72a1-4fed-9ca8-edc041e612c3","Type":"ContainerStarted","Data":"7894ea31a3b00a9592e188ae38f7f2cf9f4fb045f9a02a8573abc47504a57b7c"} Dec 06 06:06:09 crc kubenswrapper[4957]: I1206 06:06:09.051517 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:06:09 crc kubenswrapper[4957]: I1206 06:06:09.057151 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 06:06:09 crc kubenswrapper[4957]: I1206 06:06:09.059163 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:06:09 crc kubenswrapper[4957]: I1206 06:06:09.662945 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:06:09 crc kubenswrapper[4957]: E1206 06:06:09.663754 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:06:09 crc kubenswrapper[4957]: I1206 06:06:09.936402 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"963f3a5a-72a1-4fed-9ca8-edc041e612c3","Type":"ContainerStarted","Data":"75e0d3b2189b26e628cebb64b77a19637a44688c8b48d8c189b549faf91364ad"} Dec 06 06:06:09 crc kubenswrapper[4957]: I1206 06:06:09.942957 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 06:06:10 crc kubenswrapper[4957]: I1206 06:06:10.950140 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"963f3a5a-72a1-4fed-9ca8-edc041e612c3","Type":"ContainerStarted","Data":"b5bd77af2649ff90b5e0436658332049a2c6da6c2af58888c0d179fe43f13e8b"} Dec 06 06:06:10 crc kubenswrapper[4957]: I1206 06:06:10.992090 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.4839800300000001 podStartE2EDuration="4.992067812s" podCreationTimestamp="2025-12-06 06:06:06 +0000 UTC" firstStartedPulling="2025-12-06 06:06:07.065024983 +0000 UTC m=+1606.715292615" lastFinishedPulling="2025-12-06 06:06:10.573112775 +0000 UTC m=+1610.223380397" observedRunningTime="2025-12-06 06:06:10.983910679 +0000 UTC m=+1610.634178341" watchObservedRunningTime="2025-12-06 06:06:10.992067812 +0000 UTC m=+1610.642335444" Dec 06 06:06:11 crc kubenswrapper[4957]: I1206 06:06:11.958699 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 06:06:23 crc kubenswrapper[4957]: I1206 06:06:23.662413 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:06:23 crc kubenswrapper[4957]: E1206 06:06:23.663193 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:06:36 crc kubenswrapper[4957]: I1206 06:06:36.598671 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 06:06:38 crc kubenswrapper[4957]: I1206 06:06:38.666583 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:06:38 crc kubenswrapper[4957]: E1206 06:06:38.667486 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:06:47 crc kubenswrapper[4957]: I1206 06:06:47.197455 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:06:48 crc kubenswrapper[4957]: I1206 06:06:48.067292 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:06:49 crc kubenswrapper[4957]: I1206 06:06:49.662282 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:06:49 crc kubenswrapper[4957]: E1206 06:06:49.662786 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:06:51 crc kubenswrapper[4957]: I1206 06:06:51.452929 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerName="rabbitmq" containerID="cri-o://b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df" gracePeriod=604796 Dec 06 06:06:52 crc kubenswrapper[4957]: I1206 06:06:52.223601 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="rabbitmq" containerID="cri-o://cbbbd9c5a84f18728f60f9cb9c74c08a957e41c603243d9a8d1717df9c96dfe4" gracePeriod=604796 Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.101908 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227140 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-tls\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227193 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-erlang-cookie-secret\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227240 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227262 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-config-data\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227309 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-plugins\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227329 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-server-conf\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227361 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-confd\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227379 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-plugins-conf\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227403 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng6nc\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-kube-api-access-ng6nc\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227425 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-erlang-cookie\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.227455 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-pod-info\") pod \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\" (UID: \"af1ac671-ce48-4e6e-a1d6-4b7cbd079050\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.228092 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.230471 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.231198 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.237423 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.241294 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.241887 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-pod-info" (OuterVolumeSpecName: "pod-info") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.254524 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-kube-api-access-ng6nc" (OuterVolumeSpecName: "kube-api-access-ng6nc") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "kube-api-access-ng6nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.257432 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-config-data" (OuterVolumeSpecName: "config-data") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.270591 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.277917 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-server-conf" (OuterVolumeSpecName: "server-conf") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329582 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329611 4957 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329640 4957 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329650 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329659 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329668 4957 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329676 4957 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329686 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng6nc\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-kube-api-access-ng6nc\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329697 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.329706 4957 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.361494 4957 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.384134 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "af1ac671-ce48-4e6e-a1d6-4b7cbd079050" (UID: "af1ac671-ce48-4e6e-a1d6-4b7cbd079050"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.434185 4957 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.434224 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af1ac671-ce48-4e6e-a1d6-4b7cbd079050-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.657041 4957 generic.go:334] "Generic (PLEG): container finished" podID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerID="b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df" exitCode=0 Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.657129 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af1ac671-ce48-4e6e-a1d6-4b7cbd079050","Type":"ContainerDied","Data":"b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df"} Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.657157 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"af1ac671-ce48-4e6e-a1d6-4b7cbd079050","Type":"ContainerDied","Data":"dc115a275194cedd37444c709e146a1930050199882fc44fe17fbed923390369"} Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.657145 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.657176 4957 scope.go:117] "RemoveContainer" containerID="b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.666139 4957 generic.go:334] "Generic (PLEG): container finished" podID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerID="cbbbd9c5a84f18728f60f9cb9c74c08a957e41c603243d9a8d1717df9c96dfe4" exitCode=0 Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.693367 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"91bdbaa2-3bc1-44b8-b326-8feb00dc400a","Type":"ContainerDied","Data":"cbbbd9c5a84f18728f60f9cb9c74c08a957e41c603243d9a8d1717df9c96dfe4"} Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.695682 4957 scope.go:117] "RemoveContainer" containerID="ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.807786 4957 scope.go:117] "RemoveContainer" containerID="b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df" Dec 06 06:06:58 crc kubenswrapper[4957]: E1206 06:06:58.808303 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df\": container with ID starting with b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df not found: ID does not exist" containerID="b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.808355 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df"} err="failed to get container status \"b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df\": rpc error: code = NotFound desc = could not find container \"b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df\": container with ID starting with b5bbd61cd985d191e342184fb2e3aff4431dc02c88bb4a1a73efcda45b4c19df not found: ID does not exist" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.808392 4957 scope.go:117] "RemoveContainer" containerID="ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72" Dec 06 06:06:58 crc kubenswrapper[4957]: E1206 06:06:58.808721 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72\": container with ID starting with ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72 not found: ID does not exist" containerID="ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.808750 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72"} err="failed to get container status \"ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72\": rpc error: code = NotFound desc = could not find container \"ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72\": container with ID starting with ae4a8fc2f823a7027272adae2f273c77f96f53fe8f35b289b72583de7c27fc72 not found: ID does not exist" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.832538 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.945862 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-config-data\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.946283 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-tls\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.946312 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fd5c\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-kube-api-access-8fd5c\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.946374 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-erlang-cookie\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.946438 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-plugins\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.946513 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-server-conf\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.946900 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.946990 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.947009 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-pod-info\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.947141 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-plugins-conf\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.947190 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-confd\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.947267 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-erlang-cookie-secret\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.947311 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\" (UID: \"91bdbaa2-3bc1-44b8-b326-8feb00dc400a\") " Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.947622 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.948061 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.948081 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.948093 4957 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.952134 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.952645 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-pod-info" (OuterVolumeSpecName: "pod-info") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.965461 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.966103 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:58 crc kubenswrapper[4957]: I1206 06:06:58.969672 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-kube-api-access-8fd5c" (OuterVolumeSpecName: "kube-api-access-8fd5c") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "kube-api-access-8fd5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.011889 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-config-data" (OuterVolumeSpecName: "config-data") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.016178 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-server-conf" (OuterVolumeSpecName: "server-conf") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.050494 4957 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.050548 4957 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.050562 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.050576 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.050588 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fd5c\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-kube-api-access-8fd5c\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.050600 4957 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.050611 4957 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.075509 4957 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.084108 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "91bdbaa2-3bc1-44b8-b326-8feb00dc400a" (UID: "91bdbaa2-3bc1-44b8-b326-8feb00dc400a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.152522 4957 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/91bdbaa2-3bc1-44b8-b326-8feb00dc400a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.152560 4957 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.675796 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"91bdbaa2-3bc1-44b8-b326-8feb00dc400a","Type":"ContainerDied","Data":"7f8238c9029a5057464044f35695da1be88f14aac5ec6a85c246052f119b465b"} Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.675859 4957 scope.go:117] "RemoveContainer" containerID="cbbbd9c5a84f18728f60f9cb9c74c08a957e41c603243d9a8d1717df9c96dfe4" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.675946 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.751246 4957 scope.go:117] "RemoveContainer" containerID="652119ddca1221ec06b591256f24db2bbd49ae927a7b0a92db1879b283119a69" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.751810 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.760987 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.779211 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:06:59 crc kubenswrapper[4957]: E1206 06:06:59.779850 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="setup-container" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.779866 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="setup-container" Dec 06 06:06:59 crc kubenswrapper[4957]: E1206 06:06:59.779880 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerName="setup-container" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.779886 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerName="setup-container" Dec 06 06:06:59 crc kubenswrapper[4957]: E1206 06:06:59.779916 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerName="rabbitmq" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.779922 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerName="rabbitmq" Dec 06 06:06:59 crc kubenswrapper[4957]: E1206 06:06:59.779932 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="rabbitmq" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.779938 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="rabbitmq" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.780097 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="rabbitmq" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.780108 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" containerName="rabbitmq" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.781004 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.787906 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.788544 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.789046 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.789245 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.789432 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6vrsv" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.789457 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.794400 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.797484 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867106 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867166 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867195 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867222 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867274 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a28833e8-d161-4dd0-816e-b476ab8d473b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867302 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfqzr\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-kube-api-access-dfqzr\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867323 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867343 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a28833e8-d161-4dd0-816e-b476ab8d473b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867360 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867407 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.867425 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969351 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969452 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969505 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969542 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969583 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969652 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a28833e8-d161-4dd0-816e-b476ab8d473b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969697 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfqzr\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-kube-api-access-dfqzr\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969733 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969768 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a28833e8-d161-4dd0-816e-b476ab8d473b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969798 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969875 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.969903 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.970669 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.971738 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.972501 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.973055 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.974447 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a28833e8-d161-4dd0-816e-b476ab8d473b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.974892 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.978209 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a28833e8-d161-4dd0-816e-b476ab8d473b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.981248 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a28833e8-d161-4dd0-816e-b476ab8d473b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.981711 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:06:59 crc kubenswrapper[4957]: I1206 06:06:59.991037 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfqzr\" (UniqueName: \"kubernetes.io/projected/a28833e8-d161-4dd0-816e-b476ab8d473b-kube-api-access-dfqzr\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:07:00 crc kubenswrapper[4957]: I1206 06:07:00.014765 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a28833e8-d161-4dd0-816e-b476ab8d473b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:07:00 crc kubenswrapper[4957]: I1206 06:07:00.135181 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:07:01 crc kubenswrapper[4957]: I1206 06:07:00.692910 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" path="/var/lib/kubelet/pods/91bdbaa2-3bc1-44b8-b326-8feb00dc400a/volumes" Dec 06 06:07:01 crc kubenswrapper[4957]: I1206 06:07:00.698611 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a28833e8-d161-4dd0-816e-b476ab8d473b","Type":"ContainerStarted","Data":"1c4795245bf05c57e7121c5bb15b40358cd9effb9e7015f3a00cb41030cd0dc4"} Dec 06 06:07:01 crc kubenswrapper[4957]: I1206 06:07:00.698660 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 06:07:01 crc kubenswrapper[4957]: I1206 06:07:01.945932 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qc7sk"] Dec 06 06:07:01 crc kubenswrapper[4957]: I1206 06:07:01.947946 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:01 crc kubenswrapper[4957]: I1206 06:07:01.953975 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 06 06:07:01 crc kubenswrapper[4957]: I1206 06:07:01.962553 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qc7sk"] Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.058738 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.059016 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.059098 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.059206 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.059331 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-config\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.059405 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.059503 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j94mj\" (UniqueName: \"kubernetes.io/projected/43d9f52f-3f36-43d6-ab75-ff7b45650667-kube-api-access-j94mj\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.162011 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-config\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.162070 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.162159 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j94mj\" (UniqueName: \"kubernetes.io/projected/43d9f52f-3f36-43d6-ab75-ff7b45650667-kube-api-access-j94mj\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.162280 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.162378 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.162428 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.162458 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.163590 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.163702 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.163990 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.164038 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.164056 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.164235 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-config\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.188325 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j94mj\" (UniqueName: \"kubernetes.io/projected/43d9f52f-3f36-43d6-ab75-ff7b45650667-kube-api-access-j94mj\") pod \"dnsmasq-dns-79bd4cc8c9-qc7sk\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.281474 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.663819 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:07:02 crc kubenswrapper[4957]: E1206 06:07:02.664519 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.719585 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a28833e8-d161-4dd0-816e-b476ab8d473b","Type":"ContainerStarted","Data":"0933debb0f60553348194eb1bd1dd859c9fd0f9acbd38e38a7a940319e240099"} Dec 06 06:07:02 crc kubenswrapper[4957]: W1206 06:07:02.796756 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43d9f52f_3f36_43d6_ab75_ff7b45650667.slice/crio-a33df95e19366278a05358ab07cc41cb9365b1601f334ef675cc7758ca8e10c9 WatchSource:0}: Error finding container a33df95e19366278a05358ab07cc41cb9365b1601f334ef675cc7758ca8e10c9: Status 404 returned error can't find the container with id a33df95e19366278a05358ab07cc41cb9365b1601f334ef675cc7758ca8e10c9 Dec 06 06:07:02 crc kubenswrapper[4957]: I1206 06:07:02.819920 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qc7sk"] Dec 06 06:07:03 crc kubenswrapper[4957]: I1206 06:07:03.654059 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="91bdbaa2-3bc1-44b8-b326-8feb00dc400a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: i/o timeout" Dec 06 06:07:03 crc kubenswrapper[4957]: I1206 06:07:03.729298 4957 generic.go:334] "Generic (PLEG): container finished" podID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerID="a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822" exitCode=0 Dec 06 06:07:03 crc kubenswrapper[4957]: I1206 06:07:03.730063 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" event={"ID":"43d9f52f-3f36-43d6-ab75-ff7b45650667","Type":"ContainerDied","Data":"a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822"} Dec 06 06:07:03 crc kubenswrapper[4957]: I1206 06:07:03.730106 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" event={"ID":"43d9f52f-3f36-43d6-ab75-ff7b45650667","Type":"ContainerStarted","Data":"a33df95e19366278a05358ab07cc41cb9365b1601f334ef675cc7758ca8e10c9"} Dec 06 06:07:04 crc kubenswrapper[4957]: I1206 06:07:04.741822 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" event={"ID":"43d9f52f-3f36-43d6-ab75-ff7b45650667","Type":"ContainerStarted","Data":"4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb"} Dec 06 06:07:04 crc kubenswrapper[4957]: I1206 06:07:04.742392 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:04 crc kubenswrapper[4957]: I1206 06:07:04.786792 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" podStartSLOduration=3.786755738 podStartE2EDuration="3.786755738s" podCreationTimestamp="2025-12-06 06:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:07:04.774934255 +0000 UTC m=+1664.425201987" watchObservedRunningTime="2025-12-06 06:07:04.786755738 +0000 UTC m=+1664.437023420" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.284435 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.373150 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vlnfn"] Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.373551 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" podUID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerName="dnsmasq-dns" containerID="cri-o://4daee8ec0c6f6b263e156056c46796dfe9e76dc423527041ed2301d6695cac09" gracePeriod=10 Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.540015 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-7vzqk"] Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.542064 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.562210 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-7vzqk"] Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.694899 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-config\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.694969 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.694995 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.695194 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.695303 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.695372 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-dns-svc\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.695494 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96mlv\" (UniqueName: \"kubernetes.io/projected/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-kube-api-access-96mlv\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.797215 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.797319 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-dns-svc\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.797413 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96mlv\" (UniqueName: \"kubernetes.io/projected/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-kube-api-access-96mlv\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.797510 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-config\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.797572 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.797615 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.797717 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.800245 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-config\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.800706 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-dns-svc\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.800782 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.800887 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.801082 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.801189 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.827301 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96mlv\" (UniqueName: \"kubernetes.io/projected/3bac0712-aa19-4f2f-9645-eecdf3d19c5f-kube-api-access-96mlv\") pod \"dnsmasq-dns-55478c4467-7vzqk\" (UID: \"3bac0712-aa19-4f2f-9645-eecdf3d19c5f\") " pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.836108 4957 generic.go:334] "Generic (PLEG): container finished" podID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerID="4daee8ec0c6f6b263e156056c46796dfe9e76dc423527041ed2301d6695cac09" exitCode=0 Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.836175 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" event={"ID":"a2ac66e5-0300-49e5-9193-466fd0520dd6","Type":"ContainerDied","Data":"4daee8ec0c6f6b263e156056c46796dfe9e76dc423527041ed2301d6695cac09"} Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.836546 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" event={"ID":"a2ac66e5-0300-49e5-9193-466fd0520dd6","Type":"ContainerDied","Data":"66f91fbf68a73a7ab250253d38d1520b2a1f86c240f3048b23416c6fa5acdaeb"} Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.836583 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66f91fbf68a73a7ab250253d38d1520b2a1f86c240f3048b23416c6fa5acdaeb" Dec 06 06:07:12 crc kubenswrapper[4957]: I1206 06:07:12.872639 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.002864 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.103234 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-config\") pod \"a2ac66e5-0300-49e5-9193-466fd0520dd6\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.103313 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-nb\") pod \"a2ac66e5-0300-49e5-9193-466fd0520dd6\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.103384 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-swift-storage-0\") pod \"a2ac66e5-0300-49e5-9193-466fd0520dd6\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.103491 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-svc\") pod \"a2ac66e5-0300-49e5-9193-466fd0520dd6\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.103559 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-sb\") pod \"a2ac66e5-0300-49e5-9193-466fd0520dd6\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.103630 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hxhx\" (UniqueName: \"kubernetes.io/projected/a2ac66e5-0300-49e5-9193-466fd0520dd6-kube-api-access-6hxhx\") pod \"a2ac66e5-0300-49e5-9193-466fd0520dd6\" (UID: \"a2ac66e5-0300-49e5-9193-466fd0520dd6\") " Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.111142 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2ac66e5-0300-49e5-9193-466fd0520dd6-kube-api-access-6hxhx" (OuterVolumeSpecName: "kube-api-access-6hxhx") pod "a2ac66e5-0300-49e5-9193-466fd0520dd6" (UID: "a2ac66e5-0300-49e5-9193-466fd0520dd6"). InnerVolumeSpecName "kube-api-access-6hxhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.173659 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2ac66e5-0300-49e5-9193-466fd0520dd6" (UID: "a2ac66e5-0300-49e5-9193-466fd0520dd6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.176873 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2ac66e5-0300-49e5-9193-466fd0520dd6" (UID: "a2ac66e5-0300-49e5-9193-466fd0520dd6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.178073 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a2ac66e5-0300-49e5-9193-466fd0520dd6" (UID: "a2ac66e5-0300-49e5-9193-466fd0520dd6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.179023 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2ac66e5-0300-49e5-9193-466fd0520dd6" (UID: "a2ac66e5-0300-49e5-9193-466fd0520dd6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.202236 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-config" (OuterVolumeSpecName: "config") pod "a2ac66e5-0300-49e5-9193-466fd0520dd6" (UID: "a2ac66e5-0300-49e5-9193-466fd0520dd6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.206074 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.206126 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.206141 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.206169 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.206180 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2ac66e5-0300-49e5-9193-466fd0520dd6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.206190 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hxhx\" (UniqueName: \"kubernetes.io/projected/a2ac66e5-0300-49e5-9193-466fd0520dd6-kube-api-access-6hxhx\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.388881 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-7vzqk"] Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.847449 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" event={"ID":"3bac0712-aa19-4f2f-9645-eecdf3d19c5f","Type":"ContainerStarted","Data":"fc6935143d0e19ebe6d3d32a9dd990c6cf5fce702af45a8172436f19924f8def"} Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.847478 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-vlnfn" Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.885505 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vlnfn"] Dec 06 06:07:13 crc kubenswrapper[4957]: I1206 06:07:13.894091 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-vlnfn"] Dec 06 06:07:14 crc kubenswrapper[4957]: I1206 06:07:14.679301 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2ac66e5-0300-49e5-9193-466fd0520dd6" path="/var/lib/kubelet/pods/a2ac66e5-0300-49e5-9193-466fd0520dd6/volumes" Dec 06 06:07:14 crc kubenswrapper[4957]: I1206 06:07:14.858508 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" event={"ID":"3bac0712-aa19-4f2f-9645-eecdf3d19c5f","Type":"ContainerStarted","Data":"2d4ef33824feab1d8a9914147211b87e8b90751bd2a4c8dac1fc68b551970699"} Dec 06 06:07:15 crc kubenswrapper[4957]: I1206 06:07:15.870682 4957 generic.go:334] "Generic (PLEG): container finished" podID="3bac0712-aa19-4f2f-9645-eecdf3d19c5f" containerID="2d4ef33824feab1d8a9914147211b87e8b90751bd2a4c8dac1fc68b551970699" exitCode=0 Dec 06 06:07:15 crc kubenswrapper[4957]: I1206 06:07:15.870750 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" event={"ID":"3bac0712-aa19-4f2f-9645-eecdf3d19c5f","Type":"ContainerDied","Data":"2d4ef33824feab1d8a9914147211b87e8b90751bd2a4c8dac1fc68b551970699"} Dec 06 06:07:16 crc kubenswrapper[4957]: I1206 06:07:16.880788 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" event={"ID":"3bac0712-aa19-4f2f-9645-eecdf3d19c5f","Type":"ContainerStarted","Data":"05e8781e5ecdb325e05b7c6b9e914f388b943d1ebbc3bfe3cc78914097f37f45"} Dec 06 06:07:16 crc kubenswrapper[4957]: I1206 06:07:16.882079 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:16 crc kubenswrapper[4957]: I1206 06:07:16.907689 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" podStartSLOduration=4.907669659 podStartE2EDuration="4.907669659s" podCreationTimestamp="2025-12-06 06:07:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:07:16.900130034 +0000 UTC m=+1676.550397676" watchObservedRunningTime="2025-12-06 06:07:16.907669659 +0000 UTC m=+1676.557937291" Dec 06 06:07:17 crc kubenswrapper[4957]: I1206 06:07:17.663730 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:07:17 crc kubenswrapper[4957]: E1206 06:07:17.664289 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:07:22 crc kubenswrapper[4957]: I1206 06:07:22.874752 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-7vzqk" Dec 06 06:07:22 crc kubenswrapper[4957]: I1206 06:07:22.969048 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qc7sk"] Dec 06 06:07:22 crc kubenswrapper[4957]: I1206 06:07:22.969463 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" podUID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerName="dnsmasq-dns" containerID="cri-o://4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb" gracePeriod=10 Dec 06 06:07:23 crc kubenswrapper[4957]: I1206 06:07:23.957937 4957 generic.go:334] "Generic (PLEG): container finished" podID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerID="4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb" exitCode=0 Dec 06 06:07:23 crc kubenswrapper[4957]: I1206 06:07:23.957990 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:23 crc kubenswrapper[4957]: I1206 06:07:23.958043 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" event={"ID":"43d9f52f-3f36-43d6-ab75-ff7b45650667","Type":"ContainerDied","Data":"4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb"} Dec 06 06:07:23 crc kubenswrapper[4957]: I1206 06:07:23.958965 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" event={"ID":"43d9f52f-3f36-43d6-ab75-ff7b45650667","Type":"ContainerDied","Data":"a33df95e19366278a05358ab07cc41cb9365b1601f334ef675cc7758ca8e10c9"} Dec 06 06:07:23 crc kubenswrapper[4957]: I1206 06:07:23.959013 4957 scope.go:117] "RemoveContainer" containerID="4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.000216 4957 scope.go:117] "RemoveContainer" containerID="a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.023621 4957 scope.go:117] "RemoveContainer" containerID="4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb" Dec 06 06:07:24 crc kubenswrapper[4957]: E1206 06:07:24.024064 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb\": container with ID starting with 4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb not found: ID does not exist" containerID="4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.024105 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb"} err="failed to get container status \"4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb\": rpc error: code = NotFound desc = could not find container \"4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb\": container with ID starting with 4ff5f02482b1e18dfcb4da89e23d60ca369614a60359da78467daab7ffaab6bb not found: ID does not exist" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.024124 4957 scope.go:117] "RemoveContainer" containerID="a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822" Dec 06 06:07:24 crc kubenswrapper[4957]: E1206 06:07:24.024589 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822\": container with ID starting with a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822 not found: ID does not exist" containerID="a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.024611 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822"} err="failed to get container status \"a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822\": rpc error: code = NotFound desc = could not find container \"a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822\": container with ID starting with a9104a94c5f08047d2cb0617c9c86f4459f809da1fc62706711a64f5fa11b822 not found: ID does not exist" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.042012 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-swift-storage-0\") pod \"43d9f52f-3f36-43d6-ab75-ff7b45650667\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.042109 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-config\") pod \"43d9f52f-3f36-43d6-ab75-ff7b45650667\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.042160 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-nb\") pod \"43d9f52f-3f36-43d6-ab75-ff7b45650667\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.042193 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-svc\") pod \"43d9f52f-3f36-43d6-ab75-ff7b45650667\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.042244 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-sb\") pod \"43d9f52f-3f36-43d6-ab75-ff7b45650667\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.042299 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j94mj\" (UniqueName: \"kubernetes.io/projected/43d9f52f-3f36-43d6-ab75-ff7b45650667-kube-api-access-j94mj\") pod \"43d9f52f-3f36-43d6-ab75-ff7b45650667\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.042362 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-openstack-edpm-ipam\") pod \"43d9f52f-3f36-43d6-ab75-ff7b45650667\" (UID: \"43d9f52f-3f36-43d6-ab75-ff7b45650667\") " Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.051461 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43d9f52f-3f36-43d6-ab75-ff7b45650667-kube-api-access-j94mj" (OuterVolumeSpecName: "kube-api-access-j94mj") pod "43d9f52f-3f36-43d6-ab75-ff7b45650667" (UID: "43d9f52f-3f36-43d6-ab75-ff7b45650667"). InnerVolumeSpecName "kube-api-access-j94mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.097113 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "43d9f52f-3f36-43d6-ab75-ff7b45650667" (UID: "43d9f52f-3f36-43d6-ab75-ff7b45650667"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.104075 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-config" (OuterVolumeSpecName: "config") pod "43d9f52f-3f36-43d6-ab75-ff7b45650667" (UID: "43d9f52f-3f36-43d6-ab75-ff7b45650667"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.104884 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "43d9f52f-3f36-43d6-ab75-ff7b45650667" (UID: "43d9f52f-3f36-43d6-ab75-ff7b45650667"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.105968 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43d9f52f-3f36-43d6-ab75-ff7b45650667" (UID: "43d9f52f-3f36-43d6-ab75-ff7b45650667"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.106194 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "43d9f52f-3f36-43d6-ab75-ff7b45650667" (UID: "43d9f52f-3f36-43d6-ab75-ff7b45650667"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.117036 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "43d9f52f-3f36-43d6-ab75-ff7b45650667" (UID: "43d9f52f-3f36-43d6-ab75-ff7b45650667"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.145469 4957 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.145505 4957 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.145515 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.145524 4957 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.145532 4957 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.145540 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j94mj\" (UniqueName: \"kubernetes.io/projected/43d9f52f-3f36-43d6-ab75-ff7b45650667-kube-api-access-j94mj\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.145549 4957 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/43d9f52f-3f36-43d6-ab75-ff7b45650667-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.970040 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qc7sk" Dec 06 06:07:24 crc kubenswrapper[4957]: I1206 06:07:24.998727 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qc7sk"] Dec 06 06:07:25 crc kubenswrapper[4957]: I1206 06:07:25.008533 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qc7sk"] Dec 06 06:07:26 crc kubenswrapper[4957]: I1206 06:07:26.676990 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43d9f52f-3f36-43d6-ab75-ff7b45650667" path="/var/lib/kubelet/pods/43d9f52f-3f36-43d6-ab75-ff7b45650667/volumes" Dec 06 06:07:28 crc kubenswrapper[4957]: I1206 06:07:28.775359 4957 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","podaf1ac671-ce48-4e6e-a1d6-4b7cbd079050"] err="unable to destroy cgroup paths for cgroup [kubepods burstable podaf1ac671-ce48-4e6e-a1d6-4b7cbd079050] : Timed out while waiting for systemd to remove kubepods-burstable-podaf1ac671_ce48_4e6e_a1d6_4b7cbd079050.slice" Dec 06 06:07:28 crc kubenswrapper[4957]: E1206 06:07:28.775742 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable podaf1ac671-ce48-4e6e-a1d6-4b7cbd079050] : unable to destroy cgroup paths for cgroup [kubepods burstable podaf1ac671-ce48-4e6e-a1d6-4b7cbd079050] : Timed out while waiting for systemd to remove kubepods-burstable-podaf1ac671_ce48_4e6e_a1d6_4b7cbd079050.slice" pod="openstack/rabbitmq-server-0" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.006708 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.037372 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.054109 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.066122 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:07:29 crc kubenswrapper[4957]: E1206 06:07:29.066616 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerName="init" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.066642 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerName="init" Dec 06 06:07:29 crc kubenswrapper[4957]: E1206 06:07:29.066696 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerName="dnsmasq-dns" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.066705 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerName="dnsmasq-dns" Dec 06 06:07:29 crc kubenswrapper[4957]: E1206 06:07:29.066734 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerName="dnsmasq-dns" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.066743 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerName="dnsmasq-dns" Dec 06 06:07:29 crc kubenswrapper[4957]: E1206 06:07:29.066759 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerName="init" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.066768 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerName="init" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.066994 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ac66e5-0300-49e5-9193-466fd0520dd6" containerName="dnsmasq-dns" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.067032 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d9f52f-3f36-43d6-ab75-ff7b45650667" containerName="dnsmasq-dns" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.068335 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.072856 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.072954 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.072970 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.073024 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ht8kt" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.073086 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.073204 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.073324 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.081678 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.166554 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8504457c-7804-41fc-ac00-c2583a75137b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.166853 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8504457c-7804-41fc-ac00-c2583a75137b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.166885 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.166911 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.166947 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.166997 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5lv\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-kube-api-access-xm5lv\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.167030 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.167123 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.167147 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-config-data\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.167176 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.167200 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.268746 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.268878 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5lv\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-kube-api-access-xm5lv\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.268939 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269001 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269039 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-config-data\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269082 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269125 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269214 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8504457c-7804-41fc-ac00-c2583a75137b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269267 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8504457c-7804-41fc-ac00-c2583a75137b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269306 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269371 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.269757 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.270230 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-config-data\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.270253 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.270618 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.270886 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.271258 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8504457c-7804-41fc-ac00-c2583a75137b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.281689 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8504457c-7804-41fc-ac00-c2583a75137b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.282214 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.282329 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8504457c-7804-41fc-ac00-c2583a75137b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.282407 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.284941 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5lv\" (UniqueName: \"kubernetes.io/projected/8504457c-7804-41fc-ac00-c2583a75137b-kube-api-access-xm5lv\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.314463 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"8504457c-7804-41fc-ac00-c2583a75137b\") " pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.400214 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 06:07:29 crc kubenswrapper[4957]: I1206 06:07:29.859875 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 06:07:30 crc kubenswrapper[4957]: I1206 06:07:30.017533 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8504457c-7804-41fc-ac00-c2583a75137b","Type":"ContainerStarted","Data":"d5191f9458586d4af204c5bf5c1f7e16fe7de6e1e26843141db0de9c0bd1b77a"} Dec 06 06:07:30 crc kubenswrapper[4957]: I1206 06:07:30.671631 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:07:30 crc kubenswrapper[4957]: E1206 06:07:30.672018 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:07:30 crc kubenswrapper[4957]: I1206 06:07:30.674101 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af1ac671-ce48-4e6e-a1d6-4b7cbd079050" path="/var/lib/kubelet/pods/af1ac671-ce48-4e6e-a1d6-4b7cbd079050/volumes" Dec 06 06:07:32 crc kubenswrapper[4957]: I1206 06:07:32.043571 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8504457c-7804-41fc-ac00-c2583a75137b","Type":"ContainerStarted","Data":"573e61b29d74f463bfa87a26ec00a8b0867f8ec7adb8d7f93d5e7fbb6ce0f9cc"} Dec 06 06:07:35 crc kubenswrapper[4957]: I1206 06:07:35.070614 4957 generic.go:334] "Generic (PLEG): container finished" podID="a28833e8-d161-4dd0-816e-b476ab8d473b" containerID="0933debb0f60553348194eb1bd1dd859c9fd0f9acbd38e38a7a940319e240099" exitCode=0 Dec 06 06:07:35 crc kubenswrapper[4957]: I1206 06:07:35.070711 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a28833e8-d161-4dd0-816e-b476ab8d473b","Type":"ContainerDied","Data":"0933debb0f60553348194eb1bd1dd859c9fd0f9acbd38e38a7a940319e240099"} Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.054549 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982"] Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.057484 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.064274 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.067754 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.068404 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.068769 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.073407 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982"] Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.081121 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a28833e8-d161-4dd0-816e-b476ab8d473b","Type":"ContainerStarted","Data":"a394459c90444e67303562747032e693c25572da4d455c0424599643512e3d60"} Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.083079 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.103991 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz2rv\" (UniqueName: \"kubernetes.io/projected/a75e174a-7ba2-44d3-9d5d-060259173a48-kube-api-access-hz2rv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.104061 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.104100 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.104266 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.124893 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.124871921 podStartE2EDuration="37.124871921s" podCreationTimestamp="2025-12-06 06:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:07:36.123926045 +0000 UTC m=+1695.774193717" watchObservedRunningTime="2025-12-06 06:07:36.124871921 +0000 UTC m=+1695.775139603" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.205496 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz2rv\" (UniqueName: \"kubernetes.io/projected/a75e174a-7ba2-44d3-9d5d-060259173a48-kube-api-access-hz2rv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.205582 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.205612 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.205714 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.210352 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.210490 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.212441 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.226035 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz2rv\" (UniqueName: \"kubernetes.io/projected/a75e174a-7ba2-44d3-9d5d-060259173a48-kube-api-access-hz2rv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jt982\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.388080 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:07:36 crc kubenswrapper[4957]: I1206 06:07:36.980828 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982"] Dec 06 06:07:37 crc kubenswrapper[4957]: I1206 06:07:37.090769 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" event={"ID":"a75e174a-7ba2-44d3-9d5d-060259173a48","Type":"ContainerStarted","Data":"e1b61900fc62c7f0339b8c9dd4e4789c2107776a631f08388a2b037889a4899f"} Dec 06 06:07:40 crc kubenswrapper[4957]: I1206 06:07:40.807770 4957 scope.go:117] "RemoveContainer" containerID="923df90cc3dd9392e73c7f54746e648ac69f436f1ab2bb03747077e0bf10a116" Dec 06 06:07:40 crc kubenswrapper[4957]: I1206 06:07:40.844282 4957 scope.go:117] "RemoveContainer" containerID="d1c2d1a7ea89d67ac386c75749e9ba1345192477f9fcdc631cd84f1472ae331f" Dec 06 06:07:44 crc kubenswrapper[4957]: I1206 06:07:44.662444 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:07:44 crc kubenswrapper[4957]: E1206 06:07:44.663364 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:07:50 crc kubenswrapper[4957]: I1206 06:07:50.138208 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 06:07:50 crc kubenswrapper[4957]: I1206 06:07:50.233171 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" event={"ID":"a75e174a-7ba2-44d3-9d5d-060259173a48","Type":"ContainerStarted","Data":"8ce060d0aab3e33ecb71b923088222db7ffe3d51692ddc71fa11650ff65427b4"} Dec 06 06:07:50 crc kubenswrapper[4957]: I1206 06:07:50.265514 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" podStartSLOduration=2.272330296 podStartE2EDuration="14.265489942s" podCreationTimestamp="2025-12-06 06:07:36 +0000 UTC" firstStartedPulling="2025-12-06 06:07:36.982813523 +0000 UTC m=+1696.633081155" lastFinishedPulling="2025-12-06 06:07:48.975973139 +0000 UTC m=+1708.626240801" observedRunningTime="2025-12-06 06:07:50.253932767 +0000 UTC m=+1709.904200439" watchObservedRunningTime="2025-12-06 06:07:50.265489942 +0000 UTC m=+1709.915757574" Dec 06 06:07:55 crc kubenswrapper[4957]: I1206 06:07:55.662811 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:07:55 crc kubenswrapper[4957]: E1206 06:07:55.663673 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:08:01 crc kubenswrapper[4957]: I1206 06:08:01.336327 4957 generic.go:334] "Generic (PLEG): container finished" podID="a75e174a-7ba2-44d3-9d5d-060259173a48" containerID="8ce060d0aab3e33ecb71b923088222db7ffe3d51692ddc71fa11650ff65427b4" exitCode=0 Dec 06 06:08:01 crc kubenswrapper[4957]: I1206 06:08:01.336979 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" event={"ID":"a75e174a-7ba2-44d3-9d5d-060259173a48","Type":"ContainerDied","Data":"8ce060d0aab3e33ecb71b923088222db7ffe3d51692ddc71fa11650ff65427b4"} Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.809817 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.825765 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-repo-setup-combined-ca-bundle\") pod \"a75e174a-7ba2-44d3-9d5d-060259173a48\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.825937 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-inventory\") pod \"a75e174a-7ba2-44d3-9d5d-060259173a48\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.825979 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key\") pod \"a75e174a-7ba2-44d3-9d5d-060259173a48\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.826067 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz2rv\" (UniqueName: \"kubernetes.io/projected/a75e174a-7ba2-44d3-9d5d-060259173a48-kube-api-access-hz2rv\") pod \"a75e174a-7ba2-44d3-9d5d-060259173a48\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.832383 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a75e174a-7ba2-44d3-9d5d-060259173a48-kube-api-access-hz2rv" (OuterVolumeSpecName: "kube-api-access-hz2rv") pod "a75e174a-7ba2-44d3-9d5d-060259173a48" (UID: "a75e174a-7ba2-44d3-9d5d-060259173a48"). InnerVolumeSpecName "kube-api-access-hz2rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.832960 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a75e174a-7ba2-44d3-9d5d-060259173a48" (UID: "a75e174a-7ba2-44d3-9d5d-060259173a48"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:08:02 crc kubenswrapper[4957]: E1206 06:08:02.856454 4957 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key podName:a75e174a-7ba2-44d3-9d5d-060259173a48 nodeName:}" failed. No retries permitted until 2025-12-06 06:08:03.356425285 +0000 UTC m=+1723.006692917 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key") pod "a75e174a-7ba2-44d3-9d5d-060259173a48" (UID: "a75e174a-7ba2-44d3-9d5d-060259173a48") : error deleting /var/lib/kubelet/pods/a75e174a-7ba2-44d3-9d5d-060259173a48/volume-subpaths: remove /var/lib/kubelet/pods/a75e174a-7ba2-44d3-9d5d-060259173a48/volume-subpaths: no such file or directory Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.859697 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-inventory" (OuterVolumeSpecName: "inventory") pod "a75e174a-7ba2-44d3-9d5d-060259173a48" (UID: "a75e174a-7ba2-44d3-9d5d-060259173a48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.927919 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz2rv\" (UniqueName: \"kubernetes.io/projected/a75e174a-7ba2-44d3-9d5d-060259173a48-kube-api-access-hz2rv\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.928153 4957 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:02 crc kubenswrapper[4957]: I1206 06:08:02.928227 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.357151 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" event={"ID":"a75e174a-7ba2-44d3-9d5d-060259173a48","Type":"ContainerDied","Data":"e1b61900fc62c7f0339b8c9dd4e4789c2107776a631f08388a2b037889a4899f"} Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.357196 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1b61900fc62c7f0339b8c9dd4e4789c2107776a631f08388a2b037889a4899f" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.357233 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jt982" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.438422 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key\") pod \"a75e174a-7ba2-44d3-9d5d-060259173a48\" (UID: \"a75e174a-7ba2-44d3-9d5d-060259173a48\") " Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.444010 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a75e174a-7ba2-44d3-9d5d-060259173a48" (UID: "a75e174a-7ba2-44d3-9d5d-060259173a48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.463272 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w"] Dec 06 06:08:03 crc kubenswrapper[4957]: E1206 06:08:03.463645 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75e174a-7ba2-44d3-9d5d-060259173a48" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.463661 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75e174a-7ba2-44d3-9d5d-060259173a48" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.463862 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75e174a-7ba2-44d3-9d5d-060259173a48" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.464508 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.477673 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w"] Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.539716 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.539792 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdp5q\" (UniqueName: \"kubernetes.io/projected/6206b723-2b44-49ab-a3ec-93ce9d406ae8-kube-api-access-fdp5q\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.539938 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.540060 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a75e174a-7ba2-44d3-9d5d-060259173a48-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.641708 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.642374 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdp5q\" (UniqueName: \"kubernetes.io/projected/6206b723-2b44-49ab-a3ec-93ce9d406ae8-kube-api-access-fdp5q\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.642774 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.645813 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.646692 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.658714 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdp5q\" (UniqueName: \"kubernetes.io/projected/6206b723-2b44-49ab-a3ec-93ce9d406ae8-kube-api-access-fdp5q\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-rc87w\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:03 crc kubenswrapper[4957]: I1206 06:08:03.812622 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:04 crc kubenswrapper[4957]: I1206 06:08:04.371210 4957 generic.go:334] "Generic (PLEG): container finished" podID="8504457c-7804-41fc-ac00-c2583a75137b" containerID="573e61b29d74f463bfa87a26ec00a8b0867f8ec7adb8d7f93d5e7fbb6ce0f9cc" exitCode=0 Dec 06 06:08:04 crc kubenswrapper[4957]: I1206 06:08:04.371385 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8504457c-7804-41fc-ac00-c2583a75137b","Type":"ContainerDied","Data":"573e61b29d74f463bfa87a26ec00a8b0867f8ec7adb8d7f93d5e7fbb6ce0f9cc"} Dec 06 06:08:04 crc kubenswrapper[4957]: I1206 06:08:04.431824 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w"] Dec 06 06:08:04 crc kubenswrapper[4957]: I1206 06:08:04.443693 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:08:05 crc kubenswrapper[4957]: I1206 06:08:05.380948 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" event={"ID":"6206b723-2b44-49ab-a3ec-93ce9d406ae8","Type":"ContainerStarted","Data":"c3cda20d00f68a2d20b5e77e7d4b238b24b4db6aa49a0c5911a361baa4b28cd5"} Dec 06 06:08:05 crc kubenswrapper[4957]: I1206 06:08:05.383329 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8504457c-7804-41fc-ac00-c2583a75137b","Type":"ContainerStarted","Data":"5bc4d77e49825935e703ccb28fbf180dd2c718ed1b763110a6a66ba0ee051dac"} Dec 06 06:08:05 crc kubenswrapper[4957]: I1206 06:08:05.383576 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 06:08:05 crc kubenswrapper[4957]: I1206 06:08:05.418445 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.418423168 podStartE2EDuration="36.418423168s" podCreationTimestamp="2025-12-06 06:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:08:05.410510182 +0000 UTC m=+1725.060777824" watchObservedRunningTime="2025-12-06 06:08:05.418423168 +0000 UTC m=+1725.068690810" Dec 06 06:08:06 crc kubenswrapper[4957]: I1206 06:08:06.394068 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" event={"ID":"6206b723-2b44-49ab-a3ec-93ce9d406ae8","Type":"ContainerStarted","Data":"c5a1689d9944e491ac3d9a509e6c0ff0ee3712afc5d955e650bce765c93f9382"} Dec 06 06:08:06 crc kubenswrapper[4957]: I1206 06:08:06.426102 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" podStartSLOduration=2.731880193 podStartE2EDuration="3.426080907s" podCreationTimestamp="2025-12-06 06:08:03 +0000 UTC" firstStartedPulling="2025-12-06 06:08:04.443497 +0000 UTC m=+1724.093764632" lastFinishedPulling="2025-12-06 06:08:05.137697714 +0000 UTC m=+1724.787965346" observedRunningTime="2025-12-06 06:08:06.416174397 +0000 UTC m=+1726.066442059" watchObservedRunningTime="2025-12-06 06:08:06.426080907 +0000 UTC m=+1726.076348559" Dec 06 06:08:08 crc kubenswrapper[4957]: I1206 06:08:08.415122 4957 generic.go:334] "Generic (PLEG): container finished" podID="6206b723-2b44-49ab-a3ec-93ce9d406ae8" containerID="c5a1689d9944e491ac3d9a509e6c0ff0ee3712afc5d955e650bce765c93f9382" exitCode=0 Dec 06 06:08:08 crc kubenswrapper[4957]: I1206 06:08:08.415162 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" event={"ID":"6206b723-2b44-49ab-a3ec-93ce9d406ae8","Type":"ContainerDied","Data":"c5a1689d9944e491ac3d9a509e6c0ff0ee3712afc5d955e650bce765c93f9382"} Dec 06 06:08:09 crc kubenswrapper[4957]: I1206 06:08:09.662637 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:08:09 crc kubenswrapper[4957]: E1206 06:08:09.663141 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:08:09 crc kubenswrapper[4957]: I1206 06:08:09.833670 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:09 crc kubenswrapper[4957]: I1206 06:08:09.969168 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-ssh-key\") pod \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " Dec 06 06:08:09 crc kubenswrapper[4957]: I1206 06:08:09.970532 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-inventory\") pod \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " Dec 06 06:08:09 crc kubenswrapper[4957]: I1206 06:08:09.970721 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdp5q\" (UniqueName: \"kubernetes.io/projected/6206b723-2b44-49ab-a3ec-93ce9d406ae8-kube-api-access-fdp5q\") pod \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\" (UID: \"6206b723-2b44-49ab-a3ec-93ce9d406ae8\") " Dec 06 06:08:09 crc kubenswrapper[4957]: I1206 06:08:09.975759 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6206b723-2b44-49ab-a3ec-93ce9d406ae8-kube-api-access-fdp5q" (OuterVolumeSpecName: "kube-api-access-fdp5q") pod "6206b723-2b44-49ab-a3ec-93ce9d406ae8" (UID: "6206b723-2b44-49ab-a3ec-93ce9d406ae8"). InnerVolumeSpecName "kube-api-access-fdp5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.062747 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-inventory" (OuterVolumeSpecName: "inventory") pod "6206b723-2b44-49ab-a3ec-93ce9d406ae8" (UID: "6206b723-2b44-49ab-a3ec-93ce9d406ae8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.074125 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdp5q\" (UniqueName: \"kubernetes.io/projected/6206b723-2b44-49ab-a3ec-93ce9d406ae8-kube-api-access-fdp5q\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.074159 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.078042 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6206b723-2b44-49ab-a3ec-93ce9d406ae8" (UID: "6206b723-2b44-49ab-a3ec-93ce9d406ae8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.176107 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6206b723-2b44-49ab-a3ec-93ce9d406ae8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.436348 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" event={"ID":"6206b723-2b44-49ab-a3ec-93ce9d406ae8","Type":"ContainerDied","Data":"c3cda20d00f68a2d20b5e77e7d4b238b24b4db6aa49a0c5911a361baa4b28cd5"} Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.436393 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-rc87w" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.436401 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3cda20d00f68a2d20b5e77e7d4b238b24b4db6aa49a0c5911a361baa4b28cd5" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.521803 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9"] Dec 06 06:08:10 crc kubenswrapper[4957]: E1206 06:08:10.522419 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6206b723-2b44-49ab-a3ec-93ce9d406ae8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.522450 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6206b723-2b44-49ab-a3ec-93ce9d406ae8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.522756 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="6206b723-2b44-49ab-a3ec-93ce9d406ae8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.523785 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.525734 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.527419 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.527480 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.528260 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.535145 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9"] Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.685770 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.685903 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.686098 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d5hd\" (UniqueName: \"kubernetes.io/projected/46b05e85-6b4c-4aae-ab6b-39e026f48d49-kube-api-access-2d5hd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.686211 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.787787 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.788204 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.788245 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d5hd\" (UniqueName: \"kubernetes.io/projected/46b05e85-6b4c-4aae-ab6b-39e026f48d49-kube-api-access-2d5hd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.788293 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.793375 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.793408 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.795903 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.804999 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d5hd\" (UniqueName: \"kubernetes.io/projected/46b05e85-6b4c-4aae-ab6b-39e026f48d49-kube-api-access-2d5hd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:10 crc kubenswrapper[4957]: I1206 06:08:10.839671 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:08:11 crc kubenswrapper[4957]: I1206 06:08:11.425349 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9"] Dec 06 06:08:11 crc kubenswrapper[4957]: I1206 06:08:11.448664 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" event={"ID":"46b05e85-6b4c-4aae-ab6b-39e026f48d49","Type":"ContainerStarted","Data":"be869c2ae0a5e386b757241641d8ace9b99084f5f1a9b5d1e2949ea0f0b378ef"} Dec 06 06:08:12 crc kubenswrapper[4957]: I1206 06:08:12.461150 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" event={"ID":"46b05e85-6b4c-4aae-ab6b-39e026f48d49","Type":"ContainerStarted","Data":"e43caf829c3dfbfc96427fcd2d754e6dfbc5b655c5c3bcb29209d4622d1cd03b"} Dec 06 06:08:12 crc kubenswrapper[4957]: I1206 06:08:12.486415 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" podStartSLOduration=1.814544844 podStartE2EDuration="2.486392379s" podCreationTimestamp="2025-12-06 06:08:10 +0000 UTC" firstStartedPulling="2025-12-06 06:08:11.436155281 +0000 UTC m=+1731.086422913" lastFinishedPulling="2025-12-06 06:08:12.108002816 +0000 UTC m=+1731.758270448" observedRunningTime="2025-12-06 06:08:12.475116202 +0000 UTC m=+1732.125383844" watchObservedRunningTime="2025-12-06 06:08:12.486392379 +0000 UTC m=+1732.136660021" Dec 06 06:08:19 crc kubenswrapper[4957]: I1206 06:08:19.404795 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 06:08:24 crc kubenswrapper[4957]: I1206 06:08:24.662990 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:08:24 crc kubenswrapper[4957]: E1206 06:08:24.664467 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:08:37 crc kubenswrapper[4957]: I1206 06:08:37.662465 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:08:37 crc kubenswrapper[4957]: E1206 06:08:37.663366 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:08:42 crc kubenswrapper[4957]: I1206 06:08:42.060006 4957 scope.go:117] "RemoveContainer" containerID="f0629cc1dd119ea5eda64558a77e2893c31de9d4db835e9ce7d893477706c4e5" Dec 06 06:08:42 crc kubenswrapper[4957]: I1206 06:08:42.083569 4957 scope.go:117] "RemoveContainer" containerID="bdf3a62a177b2218ae66053a3efbaaaff377a66d1b101a5263c182572371b465" Dec 06 06:08:42 crc kubenswrapper[4957]: I1206 06:08:42.130963 4957 scope.go:117] "RemoveContainer" containerID="92947cccf0669f5e9c7584d557033334706aa9b8d23cc7787538bea6adb644fa" Dec 06 06:08:50 crc kubenswrapper[4957]: I1206 06:08:50.672005 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:08:50 crc kubenswrapper[4957]: E1206 06:08:50.672790 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:09:04 crc kubenswrapper[4957]: I1206 06:09:04.662123 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:09:04 crc kubenswrapper[4957]: E1206 06:09:04.662952 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:09:19 crc kubenswrapper[4957]: I1206 06:09:19.662996 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:09:19 crc kubenswrapper[4957]: E1206 06:09:19.665880 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:09:31 crc kubenswrapper[4957]: I1206 06:09:31.663435 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:09:31 crc kubenswrapper[4957]: E1206 06:09:31.665740 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:09:42 crc kubenswrapper[4957]: I1206 06:09:42.663072 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:09:42 crc kubenswrapper[4957]: E1206 06:09:42.664476 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.214928 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vncrz"] Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.218433 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.250719 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vncrz"] Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.312668 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-utilities\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.312744 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-catalog-content\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.312789 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdwdl\" (UniqueName: \"kubernetes.io/projected/6743a778-7911-4a8d-9052-d0cf4ba631b7-kube-api-access-jdwdl\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.415002 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-catalog-content\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.415073 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdwdl\" (UniqueName: \"kubernetes.io/projected/6743a778-7911-4a8d-9052-d0cf4ba631b7-kube-api-access-jdwdl\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.415234 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-utilities\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.415595 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-catalog-content\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.415976 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-utilities\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.434726 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdwdl\" (UniqueName: \"kubernetes.io/projected/6743a778-7911-4a8d-9052-d0cf4ba631b7-kube-api-access-jdwdl\") pod \"certified-operators-vncrz\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.592552 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:09:56 crc kubenswrapper[4957]: I1206 06:09:56.663729 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:09:56 crc kubenswrapper[4957]: E1206 06:09:56.664118 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:09:57 crc kubenswrapper[4957]: I1206 06:09:57.190514 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vncrz"] Dec 06 06:09:57 crc kubenswrapper[4957]: I1206 06:09:57.543487 4957 generic.go:334] "Generic (PLEG): container finished" podID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerID="2d4ad53caf41e4d13ec6205760164d45b18adbd8c913c21b9638dcae7572e961" exitCode=0 Dec 06 06:09:57 crc kubenswrapper[4957]: I1206 06:09:57.543528 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vncrz" event={"ID":"6743a778-7911-4a8d-9052-d0cf4ba631b7","Type":"ContainerDied","Data":"2d4ad53caf41e4d13ec6205760164d45b18adbd8c913c21b9638dcae7572e961"} Dec 06 06:09:57 crc kubenswrapper[4957]: I1206 06:09:57.543551 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vncrz" event={"ID":"6743a778-7911-4a8d-9052-d0cf4ba631b7","Type":"ContainerStarted","Data":"fa41111525f88b6928d2d0c0f40e79ab84f0b26d8ec65461fafe19df03320a95"} Dec 06 06:10:00 crc kubenswrapper[4957]: I1206 06:10:00.580007 4957 generic.go:334] "Generic (PLEG): container finished" podID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerID="da8507de11683694bcea84288e958c9b7192002b7e33d65ed79cb4f1fb67b67f" exitCode=0 Dec 06 06:10:00 crc kubenswrapper[4957]: I1206 06:10:00.580070 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vncrz" event={"ID":"6743a778-7911-4a8d-9052-d0cf4ba631b7","Type":"ContainerDied","Data":"da8507de11683694bcea84288e958c9b7192002b7e33d65ed79cb4f1fb67b67f"} Dec 06 06:10:02 crc kubenswrapper[4957]: I1206 06:10:02.605584 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vncrz" event={"ID":"6743a778-7911-4a8d-9052-d0cf4ba631b7","Type":"ContainerStarted","Data":"89a50759ee4d441f99c0a648e5f6b854b550faa915d0f695d33757de8a3d0f1c"} Dec 06 06:10:02 crc kubenswrapper[4957]: I1206 06:10:02.634576 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vncrz" podStartSLOduration=2.148693802 podStartE2EDuration="6.634548278s" podCreationTimestamp="2025-12-06 06:09:56 +0000 UTC" firstStartedPulling="2025-12-06 06:09:57.546865214 +0000 UTC m=+1837.197132846" lastFinishedPulling="2025-12-06 06:10:02.03271965 +0000 UTC m=+1841.682987322" observedRunningTime="2025-12-06 06:10:02.628354699 +0000 UTC m=+1842.278622341" watchObservedRunningTime="2025-12-06 06:10:02.634548278 +0000 UTC m=+1842.284815940" Dec 06 06:10:06 crc kubenswrapper[4957]: I1206 06:10:06.592869 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:10:06 crc kubenswrapper[4957]: I1206 06:10:06.595031 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:10:06 crc kubenswrapper[4957]: I1206 06:10:06.684188 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:10:08 crc kubenswrapper[4957]: I1206 06:10:08.662796 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:10:08 crc kubenswrapper[4957]: E1206 06:10:08.663429 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:10:16 crc kubenswrapper[4957]: I1206 06:10:16.679808 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:10:16 crc kubenswrapper[4957]: I1206 06:10:16.737861 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vncrz"] Dec 06 06:10:16 crc kubenswrapper[4957]: I1206 06:10:16.821879 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vncrz" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="registry-server" containerID="cri-o://89a50759ee4d441f99c0a648e5f6b854b550faa915d0f695d33757de8a3d0f1c" gracePeriod=2 Dec 06 06:10:18 crc kubenswrapper[4957]: I1206 06:10:18.856441 4957 generic.go:334] "Generic (PLEG): container finished" podID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerID="89a50759ee4d441f99c0a648e5f6b854b550faa915d0f695d33757de8a3d0f1c" exitCode=0 Dec 06 06:10:18 crc kubenswrapper[4957]: I1206 06:10:18.856509 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vncrz" event={"ID":"6743a778-7911-4a8d-9052-d0cf4ba631b7","Type":"ContainerDied","Data":"89a50759ee4d441f99c0a648e5f6b854b550faa915d0f695d33757de8a3d0f1c"} Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.225063 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.414052 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdwdl\" (UniqueName: \"kubernetes.io/projected/6743a778-7911-4a8d-9052-d0cf4ba631b7-kube-api-access-jdwdl\") pod \"6743a778-7911-4a8d-9052-d0cf4ba631b7\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.414149 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-utilities\") pod \"6743a778-7911-4a8d-9052-d0cf4ba631b7\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.414525 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-catalog-content\") pod \"6743a778-7911-4a8d-9052-d0cf4ba631b7\" (UID: \"6743a778-7911-4a8d-9052-d0cf4ba631b7\") " Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.415818 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-utilities" (OuterVolumeSpecName: "utilities") pod "6743a778-7911-4a8d-9052-d0cf4ba631b7" (UID: "6743a778-7911-4a8d-9052-d0cf4ba631b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.419891 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6743a778-7911-4a8d-9052-d0cf4ba631b7-kube-api-access-jdwdl" (OuterVolumeSpecName: "kube-api-access-jdwdl") pod "6743a778-7911-4a8d-9052-d0cf4ba631b7" (UID: "6743a778-7911-4a8d-9052-d0cf4ba631b7"). InnerVolumeSpecName "kube-api-access-jdwdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.472087 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6743a778-7911-4a8d-9052-d0cf4ba631b7" (UID: "6743a778-7911-4a8d-9052-d0cf4ba631b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.516759 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.516792 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdwdl\" (UniqueName: \"kubernetes.io/projected/6743a778-7911-4a8d-9052-d0cf4ba631b7-kube-api-access-jdwdl\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.516803 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6743a778-7911-4a8d-9052-d0cf4ba631b7-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.874546 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vncrz" event={"ID":"6743a778-7911-4a8d-9052-d0cf4ba631b7","Type":"ContainerDied","Data":"fa41111525f88b6928d2d0c0f40e79ab84f0b26d8ec65461fafe19df03320a95"} Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.874615 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vncrz" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.874641 4957 scope.go:117] "RemoveContainer" containerID="89a50759ee4d441f99c0a648e5f6b854b550faa915d0f695d33757de8a3d0f1c" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.917491 4957 scope.go:117] "RemoveContainer" containerID="da8507de11683694bcea84288e958c9b7192002b7e33d65ed79cb4f1fb67b67f" Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.931399 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vncrz"] Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.945253 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vncrz"] Dec 06 06:10:19 crc kubenswrapper[4957]: I1206 06:10:19.945777 4957 scope.go:117] "RemoveContainer" containerID="2d4ad53caf41e4d13ec6205760164d45b18adbd8c913c21b9638dcae7572e961" Dec 06 06:10:20 crc kubenswrapper[4957]: I1206 06:10:20.684553 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" path="/var/lib/kubelet/pods/6743a778-7911-4a8d-9052-d0cf4ba631b7/volumes" Dec 06 06:10:22 crc kubenswrapper[4957]: I1206 06:10:22.662608 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:10:22 crc kubenswrapper[4957]: E1206 06:10:22.662947 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:10:35 crc kubenswrapper[4957]: I1206 06:10:35.662772 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:10:35 crc kubenswrapper[4957]: E1206 06:10:35.663955 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:10:47 crc kubenswrapper[4957]: I1206 06:10:47.663329 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:10:49 crc kubenswrapper[4957]: I1206 06:10:49.194823 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"5117ce9e55733277f221b30cc29bce4ae29d818634565ff5fe20d5464771feb5"} Dec 06 06:11:11 crc kubenswrapper[4957]: I1206 06:11:11.048240 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-d882-account-create-update-kwl95"] Dec 06 06:11:11 crc kubenswrapper[4957]: I1206 06:11:11.058221 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-46vj4"] Dec 06 06:11:11 crc kubenswrapper[4957]: I1206 06:11:11.066318 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-46vj4"] Dec 06 06:11:11 crc kubenswrapper[4957]: I1206 06:11:11.074926 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-d882-account-create-update-kwl95"] Dec 06 06:11:12 crc kubenswrapper[4957]: I1206 06:11:12.674645 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c28de4c2-bcee-4913-9153-157e8e33dda0" path="/var/lib/kubelet/pods/c28de4c2-bcee-4913-9153-157e8e33dda0/volumes" Dec 06 06:11:12 crc kubenswrapper[4957]: I1206 06:11:12.677289 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2b3373d-2571-42a7-b265-ff1409d332a0" path="/var/lib/kubelet/pods/c2b3373d-2571-42a7-b265-ff1409d332a0/volumes" Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.038693 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-60fa-account-create-update-mpw77"] Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.052449 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-hhr4x"] Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.063174 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-8wrvt"] Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.074549 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-8wrvt"] Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.084400 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-60fa-account-create-update-mpw77"] Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.098556 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-hhr4x"] Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.109979 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-03d3-account-create-update-gjflt"] Dec 06 06:11:21 crc kubenswrapper[4957]: I1206 06:11:21.119954 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-03d3-account-create-update-gjflt"] Dec 06 06:11:22 crc kubenswrapper[4957]: I1206 06:11:22.679023 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9" path="/var/lib/kubelet/pods/16d14c89-f07e-4c7d-8a3f-3f7599fb7bf9/volumes" Dec 06 06:11:22 crc kubenswrapper[4957]: I1206 06:11:22.679653 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58d697c5-d777-48f0-a82e-5b685102c543" path="/var/lib/kubelet/pods/58d697c5-d777-48f0-a82e-5b685102c543/volumes" Dec 06 06:11:22 crc kubenswrapper[4957]: I1206 06:11:22.680259 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4fe16c0-afe2-4b62-b62b-108ee7b0dddb" path="/var/lib/kubelet/pods/b4fe16c0-afe2-4b62-b62b-108ee7b0dddb/volumes" Dec 06 06:11:22 crc kubenswrapper[4957]: I1206 06:11:22.680885 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3026135-ca82-4fab-a560-4cb0a64e227b" path="/var/lib/kubelet/pods/e3026135-ca82-4fab-a560-4cb0a64e227b/volumes" Dec 06 06:11:39 crc kubenswrapper[4957]: I1206 06:11:39.057084 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-scxvl"] Dec 06 06:11:39 crc kubenswrapper[4957]: I1206 06:11:39.066632 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-scxvl"] Dec 06 06:11:40 crc kubenswrapper[4957]: I1206 06:11:40.699957 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98dc1a26-e2e0-4447-a841-8507984b8862" path="/var/lib/kubelet/pods/98dc1a26-e2e0-4447-a841-8507984b8862/volumes" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.318074 4957 scope.go:117] "RemoveContainer" containerID="04bbdddda2de1167150f8aa0ebcfef25ee7ec2f1cb8f7693a928eed16299a503" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.359118 4957 scope.go:117] "RemoveContainer" containerID="b6eb501e08ec6789d09c4c9926e664db4b88841e68e31a4d04c3a3f5ff56fd4d" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.432798 4957 scope.go:117] "RemoveContainer" containerID="32d5b55b0977f5c8dde43cb01e93b73ff4fa2f60bc6967aa36c11336e7389892" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.499097 4957 scope.go:117] "RemoveContainer" containerID="7833c38e87729d4fea3cf2bfc43624395dcf36d7ef229bb2c3ff57c271cac9c4" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.534848 4957 scope.go:117] "RemoveContainer" containerID="6303e4d02345bcf49a6dfb7dea4862406d20b2007dfb8ddbcd167a45067e42e2" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.596719 4957 scope.go:117] "RemoveContainer" containerID="4c8b6674d15d7b91045bb18c121ac04a1270deda62a69aab95c9d178013e71b0" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.624481 4957 scope.go:117] "RemoveContainer" containerID="c6ab871bb415b2afe7e5583240f10e5ccc07a9811e34299e559d34a4ef6dd403" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.655904 4957 scope.go:117] "RemoveContainer" containerID="c6cb89a94d7d6e6ce03f464bc9ac4b7a8ed6e181d3c005801e90785499667870" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.719039 4957 scope.go:117] "RemoveContainer" containerID="6719e8e343360b869dca20c4e5d038934e54fd5acf1a13276a5c1cde052816f4" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.745825 4957 scope.go:117] "RemoveContainer" containerID="bccad0d088bd0fccd396fad3450b9fbb1609c862785242a8c80af425fcac1ceb" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.779276 4957 scope.go:117] "RemoveContainer" containerID="4daee8ec0c6f6b263e156056c46796dfe9e76dc423527041ed2301d6695cac09" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.808267 4957 scope.go:117] "RemoveContainer" containerID="8867c86d2b2990d2807f29679c6133960ebd025317ef5b7906227a45aed90f06" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.826024 4957 scope.go:117] "RemoveContainer" containerID="7c19bfc973175473ca1d34592555409e2a43abeda269082efb9ccbeed0ba1223" Dec 06 06:11:42 crc kubenswrapper[4957]: I1206 06:11:42.845980 4957 scope.go:117] "RemoveContainer" containerID="5f6e0499693b01d434bc35403cb87cd843b292a977574015ac0be27aa71db586" Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.055699 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-m9kbj"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.066528 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-ptnbn"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.078344 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8f93-account-create-update-4ktnk"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.088329 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-ptnbn"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.099453 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8f93-account-create-update-4ktnk"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.107063 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-4a8c-account-create-update-hdlfp"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.141548 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-82qnj"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.149616 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-82qnj"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.158557 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-m9kbj"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.167208 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-4a8c-account-create-update-hdlfp"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.175127 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3e8e-account-create-update-d5w77"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.182024 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3e8e-account-create-update-d5w77"] Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.677224 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b916a3e-348a-4b54-8c65-d93e3d684aa8" path="/var/lib/kubelet/pods/4b916a3e-348a-4b54-8c65-d93e3d684aa8/volumes" Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.678079 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc78fa5-c7af-4ad7-80aa-2d57b91382f1" path="/var/lib/kubelet/pods/7bc78fa5-c7af-4ad7-80aa-2d57b91382f1/volumes" Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.678627 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9497a720-6416-4d95-87b5-98c6afed7a4c" path="/var/lib/kubelet/pods/9497a720-6416-4d95-87b5-98c6afed7a4c/volumes" Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.679174 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a264853d-5703-430c-ac2d-79d1a1170fa6" path="/var/lib/kubelet/pods/a264853d-5703-430c-ac2d-79d1a1170fa6/volumes" Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.680163 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a963cc41-8008-4761-bf82-84b3e9ab4311" path="/var/lib/kubelet/pods/a963cc41-8008-4761-bf82-84b3e9ab4311/volumes" Dec 06 06:11:58 crc kubenswrapper[4957]: I1206 06:11:58.680705 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c491d07c-4774-4670-8467-cd1eed27fd1f" path="/var/lib/kubelet/pods/c491d07c-4774-4670-8467-cd1eed27fd1f/volumes" Dec 06 06:12:00 crc kubenswrapper[4957]: I1206 06:12:00.933782 4957 generic.go:334] "Generic (PLEG): container finished" podID="46b05e85-6b4c-4aae-ab6b-39e026f48d49" containerID="e43caf829c3dfbfc96427fcd2d754e6dfbc5b655c5c3bcb29209d4622d1cd03b" exitCode=0 Dec 06 06:12:00 crc kubenswrapper[4957]: I1206 06:12:00.933892 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" event={"ID":"46b05e85-6b4c-4aae-ab6b-39e026f48d49","Type":"ContainerDied","Data":"e43caf829c3dfbfc96427fcd2d754e6dfbc5b655c5c3bcb29209d4622d1cd03b"} Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.382800 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.468064 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-inventory\") pod \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.468187 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-bootstrap-combined-ca-bundle\") pod \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.468382 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-ssh-key\") pod \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.468417 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d5hd\" (UniqueName: \"kubernetes.io/projected/46b05e85-6b4c-4aae-ab6b-39e026f48d49-kube-api-access-2d5hd\") pod \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\" (UID: \"46b05e85-6b4c-4aae-ab6b-39e026f48d49\") " Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.473686 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "46b05e85-6b4c-4aae-ab6b-39e026f48d49" (UID: "46b05e85-6b4c-4aae-ab6b-39e026f48d49"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.474296 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b05e85-6b4c-4aae-ab6b-39e026f48d49-kube-api-access-2d5hd" (OuterVolumeSpecName: "kube-api-access-2d5hd") pod "46b05e85-6b4c-4aae-ab6b-39e026f48d49" (UID: "46b05e85-6b4c-4aae-ab6b-39e026f48d49"). InnerVolumeSpecName "kube-api-access-2d5hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.500528 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-inventory" (OuterVolumeSpecName: "inventory") pod "46b05e85-6b4c-4aae-ab6b-39e026f48d49" (UID: "46b05e85-6b4c-4aae-ab6b-39e026f48d49"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.500779 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "46b05e85-6b4c-4aae-ab6b-39e026f48d49" (UID: "46b05e85-6b4c-4aae-ab6b-39e026f48d49"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.570026 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.570057 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d5hd\" (UniqueName: \"kubernetes.io/projected/46b05e85-6b4c-4aae-ab6b-39e026f48d49-kube-api-access-2d5hd\") on node \"crc\" DevicePath \"\"" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.570068 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.570078 4957 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b05e85-6b4c-4aae-ab6b-39e026f48d49-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.958385 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" event={"ID":"46b05e85-6b4c-4aae-ab6b-39e026f48d49","Type":"ContainerDied","Data":"be869c2ae0a5e386b757241641d8ace9b99084f5f1a9b5d1e2949ea0f0b378ef"} Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.958770 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be869c2ae0a5e386b757241641d8ace9b99084f5f1a9b5d1e2949ea0f0b378ef" Dec 06 06:12:02 crc kubenswrapper[4957]: I1206 06:12:02.958658 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.056634 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7"] Dec 06 06:12:03 crc kubenswrapper[4957]: E1206 06:12:03.057355 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b05e85-6b4c-4aae-ab6b-39e026f48d49" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.057395 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b05e85-6b4c-4aae-ab6b-39e026f48d49" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 06:12:03 crc kubenswrapper[4957]: E1206 06:12:03.057436 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="extract-utilities" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.057450 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="extract-utilities" Dec 06 06:12:03 crc kubenswrapper[4957]: E1206 06:12:03.057483 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="registry-server" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.057494 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="registry-server" Dec 06 06:12:03 crc kubenswrapper[4957]: E1206 06:12:03.057513 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="extract-content" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.057524 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="extract-content" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.057972 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="6743a778-7911-4a8d-9052-d0cf4ba631b7" containerName="registry-server" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.058014 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b05e85-6b4c-4aae-ab6b-39e026f48d49" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.059109 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.063978 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.063979 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.064229 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.065438 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.074941 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7"] Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.188257 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf8ml\" (UniqueName: \"kubernetes.io/projected/2dc93721-02c2-4553-960f-71955c73b548-kube-api-access-qf8ml\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.188493 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.188634 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.290008 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.290078 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.290134 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf8ml\" (UniqueName: \"kubernetes.io/projected/2dc93721-02c2-4553-960f-71955c73b548-kube-api-access-qf8ml\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.294813 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.294873 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.306549 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf8ml\" (UniqueName: \"kubernetes.io/projected/2dc93721-02c2-4553-960f-71955c73b548-kube-api-access-qf8ml\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-shvb7\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.390096 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.955518 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7"] Dec 06 06:12:03 crc kubenswrapper[4957]: I1206 06:12:03.966508 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" event={"ID":"2dc93721-02c2-4553-960f-71955c73b548","Type":"ContainerStarted","Data":"15beaf16eb6d4c8333579c430526cdbcdf8600a933483769a7f7bf413c5c21dc"} Dec 06 06:12:04 crc kubenswrapper[4957]: I1206 06:12:04.978360 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" event={"ID":"2dc93721-02c2-4553-960f-71955c73b548","Type":"ContainerStarted","Data":"9fd873227c6102415b067af0ce0b52ee56bb2b71f7652c4277b5d48fa9909160"} Dec 06 06:12:05 crc kubenswrapper[4957]: I1206 06:12:05.004685 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" podStartSLOduration=1.523635245 podStartE2EDuration="2.004664938s" podCreationTimestamp="2025-12-06 06:12:03 +0000 UTC" firstStartedPulling="2025-12-06 06:12:03.950377337 +0000 UTC m=+1963.600644969" lastFinishedPulling="2025-12-06 06:12:04.43140702 +0000 UTC m=+1964.081674662" observedRunningTime="2025-12-06 06:12:05.000198468 +0000 UTC m=+1964.650466110" watchObservedRunningTime="2025-12-06 06:12:05.004664938 +0000 UTC m=+1964.654932570" Dec 06 06:12:08 crc kubenswrapper[4957]: I1206 06:12:08.043248 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-zp2bn"] Dec 06 06:12:08 crc kubenswrapper[4957]: I1206 06:12:08.054279 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-zp2bn"] Dec 06 06:12:08 crc kubenswrapper[4957]: I1206 06:12:08.678115 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fce1a2e-8245-49f0-a949-091a9165bbc4" path="/var/lib/kubelet/pods/7fce1a2e-8245-49f0-a949-091a9165bbc4/volumes" Dec 06 06:12:40 crc kubenswrapper[4957]: I1206 06:12:40.073091 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-gv9kk"] Dec 06 06:12:40 crc kubenswrapper[4957]: I1206 06:12:40.087976 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-gv9kk"] Dec 06 06:12:40 crc kubenswrapper[4957]: I1206 06:12:40.673766 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d90cadb8-4faf-4c1a-b198-0a6b4d537bdc" path="/var/lib/kubelet/pods/d90cadb8-4faf-4c1a-b198-0a6b4d537bdc/volumes" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.102305 4957 scope.go:117] "RemoveContainer" containerID="5d8f9614304621184fad18f16681dfdce0daa5a7979ace5e0986af77660eef17" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.140068 4957 scope.go:117] "RemoveContainer" containerID="357a6f1005dac5ae3f412795cd9ee52201c5b29114480d41bda62cc119396f97" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.189814 4957 scope.go:117] "RemoveContainer" containerID="7258bbc3a39b70bee3a5bfb780226cae5c2cee87f1e9957ae4e79ff13c77f180" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.240416 4957 scope.go:117] "RemoveContainer" containerID="51d293cba8bbe646b4a898c8fb4204635f59dad77bc0cddb72222848af2a81d4" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.277279 4957 scope.go:117] "RemoveContainer" containerID="1eb52ecca5bfcce8f4099786c7777e21f2290a1d3ffec089a7b4e1f40b8ed71c" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.625382 4957 scope.go:117] "RemoveContainer" containerID="1007a9cb5ae67115e32f98650fe48e9697435c15a8b3c87316a2bbb566c3ed85" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.660190 4957 scope.go:117] "RemoveContainer" containerID="956276e152d070d60b34172451f34b8ea8b4798743901cbc7179f04b02bcf206" Dec 06 06:12:43 crc kubenswrapper[4957]: I1206 06:12:43.702984 4957 scope.go:117] "RemoveContainer" containerID="e17712189794c76133c5e4a307894e575a83d7c3d21af732826c132673e8af98" Dec 06 06:12:54 crc kubenswrapper[4957]: I1206 06:12:54.051228 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bqh2v"] Dec 06 06:12:54 crc kubenswrapper[4957]: I1206 06:12:54.060272 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bqh2v"] Dec 06 06:12:54 crc kubenswrapper[4957]: I1206 06:12:54.681324 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8a45479-cc0d-40ce-b54d-d65ee56a6c36" path="/var/lib/kubelet/pods/e8a45479-cc0d-40ce-b54d-d65ee56a6c36/volumes" Dec 06 06:12:57 crc kubenswrapper[4957]: I1206 06:12:57.033162 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zsklh"] Dec 06 06:12:57 crc kubenswrapper[4957]: I1206 06:12:57.044055 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zsklh"] Dec 06 06:12:58 crc kubenswrapper[4957]: I1206 06:12:58.675584 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c167b57e-b623-4080-a770-d87233e00128" path="/var/lib/kubelet/pods/c167b57e-b623-4080-a770-d87233e00128/volumes" Dec 06 06:13:03 crc kubenswrapper[4957]: I1206 06:13:03.069278 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-f45dr"] Dec 06 06:13:03 crc kubenswrapper[4957]: I1206 06:13:03.081059 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-f45dr"] Dec 06 06:13:04 crc kubenswrapper[4957]: I1206 06:13:04.681306 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d3a3efb-65b9-4a90-bded-99969b537744" path="/var/lib/kubelet/pods/7d3a3efb-65b9-4a90-bded-99969b537744/volumes" Dec 06 06:13:05 crc kubenswrapper[4957]: I1206 06:13:05.040694 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-6szxq"] Dec 06 06:13:05 crc kubenswrapper[4957]: I1206 06:13:05.050370 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-6szxq"] Dec 06 06:13:06 crc kubenswrapper[4957]: I1206 06:13:06.674481 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caa7934f-ce05-4cac-b198-b67ffa3e96ed" path="/var/lib/kubelet/pods/caa7934f-ce05-4cac-b198-b67ffa3e96ed/volumes" Dec 06 06:13:10 crc kubenswrapper[4957]: I1206 06:13:10.181063 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:13:10 crc kubenswrapper[4957]: I1206 06:13:10.181513 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:13:40 crc kubenswrapper[4957]: I1206 06:13:40.180874 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:13:40 crc kubenswrapper[4957]: I1206 06:13:40.181507 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:13:43 crc kubenswrapper[4957]: I1206 06:13:43.880429 4957 scope.go:117] "RemoveContainer" containerID="b70e91a0cf14b35a501421be0dec610a6bfe3f59df20f9c4f06da3e3c089f43d" Dec 06 06:13:44 crc kubenswrapper[4957]: I1206 06:13:44.170179 4957 scope.go:117] "RemoveContainer" containerID="14b61a9a5561fe1188d263899b81f4fc8d0af35a3867fae10c46438c1746f01a" Dec 06 06:13:44 crc kubenswrapper[4957]: I1206 06:13:44.210971 4957 scope.go:117] "RemoveContainer" containerID="67c4ebc25385b97d5d798360d85799ca98f251b184a5b874655814f3da6cd083" Dec 06 06:13:44 crc kubenswrapper[4957]: I1206 06:13:44.262901 4957 scope.go:117] "RemoveContainer" containerID="cea7f2dbbfccaccec7e7a87cd579143a35051da4c050e15ff8d8782c1f013fba" Dec 06 06:13:49 crc kubenswrapper[4957]: E1206 06:13:49.692935 4957 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.031s" Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.071888 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f97e-account-create-update-vgh4p"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.083446 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-4759m"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.091805 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-sv4lc"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.098821 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8e3f-account-create-update-5vz7b"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.105933 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-e1ca-account-create-update-gr9fq"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.113079 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8e3f-account-create-update-5vz7b"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.120152 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f97e-account-create-update-vgh4p"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.128871 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-sv4lc"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.138951 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-4759m"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.146875 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-e1ca-account-create-update-gr9fq"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.153669 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xfhvr"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.161527 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xfhvr"] Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.674550 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33edcf87-18b7-4f62-892d-6ab127062c66" path="/var/lib/kubelet/pods/33edcf87-18b7-4f62-892d-6ab127062c66/volumes" Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.675674 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63851a7a-b1a9-4e6e-9260-63d2b9603a16" path="/var/lib/kubelet/pods/63851a7a-b1a9-4e6e-9260-63d2b9603a16/volumes" Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.676515 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aaf6832-b63f-418e-be87-b74fd650c210" path="/var/lib/kubelet/pods/8aaf6832-b63f-418e-be87-b74fd650c210/volumes" Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.677196 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58" path="/var/lib/kubelet/pods/8ea4d72b-9cdf-45e7-9ce5-0e1da0b85f58/volumes" Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.678869 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5080f24-5973-46b2-a7b0-e5c74a8d7e72" path="/var/lib/kubelet/pods/b5080f24-5973-46b2-a7b0-e5c74a8d7e72/volumes" Dec 06 06:13:56 crc kubenswrapper[4957]: I1206 06:13:56.679521 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35ff771-af68-49da-91a9-f02b8ffb2d6a" path="/var/lib/kubelet/pods/c35ff771-af68-49da-91a9-f02b8ffb2d6a/volumes" Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.180393 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.181245 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.181309 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.182254 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5117ce9e55733277f221b30cc29bce4ae29d818634565ff5fe20d5464771feb5"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.182313 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://5117ce9e55733277f221b30cc29bce4ae29d818634565ff5fe20d5464771feb5" gracePeriod=600 Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.939068 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="5117ce9e55733277f221b30cc29bce4ae29d818634565ff5fe20d5464771feb5" exitCode=0 Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.939491 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"5117ce9e55733277f221b30cc29bce4ae29d818634565ff5fe20d5464771feb5"} Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.939796 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092"} Dec 06 06:14:10 crc kubenswrapper[4957]: I1206 06:14:10.939841 4957 scope.go:117] "RemoveContainer" containerID="4cffa31404d985dc663c0959d7936c1936c025ba54808fe22ff47efa189d14cc" Dec 06 06:14:13 crc kubenswrapper[4957]: I1206 06:14:13.970551 4957 generic.go:334] "Generic (PLEG): container finished" podID="2dc93721-02c2-4553-960f-71955c73b548" containerID="9fd873227c6102415b067af0ce0b52ee56bb2b71f7652c4277b5d48fa9909160" exitCode=0 Dec 06 06:14:13 crc kubenswrapper[4957]: I1206 06:14:13.970658 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" event={"ID":"2dc93721-02c2-4553-960f-71955c73b548","Type":"ContainerDied","Data":"9fd873227c6102415b067af0ce0b52ee56bb2b71f7652c4277b5d48fa9909160"} Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.392094 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.496581 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf8ml\" (UniqueName: \"kubernetes.io/projected/2dc93721-02c2-4553-960f-71955c73b548-kube-api-access-qf8ml\") pod \"2dc93721-02c2-4553-960f-71955c73b548\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.496805 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-ssh-key\") pod \"2dc93721-02c2-4553-960f-71955c73b548\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.496896 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-inventory\") pod \"2dc93721-02c2-4553-960f-71955c73b548\" (UID: \"2dc93721-02c2-4553-960f-71955c73b548\") " Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.502566 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dc93721-02c2-4553-960f-71955c73b548-kube-api-access-qf8ml" (OuterVolumeSpecName: "kube-api-access-qf8ml") pod "2dc93721-02c2-4553-960f-71955c73b548" (UID: "2dc93721-02c2-4553-960f-71955c73b548"). InnerVolumeSpecName "kube-api-access-qf8ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.529639 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2dc93721-02c2-4553-960f-71955c73b548" (UID: "2dc93721-02c2-4553-960f-71955c73b548"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.539009 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-inventory" (OuterVolumeSpecName: "inventory") pod "2dc93721-02c2-4553-960f-71955c73b548" (UID: "2dc93721-02c2-4553-960f-71955c73b548"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.599474 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf8ml\" (UniqueName: \"kubernetes.io/projected/2dc93721-02c2-4553-960f-71955c73b548-kube-api-access-qf8ml\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.599780 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.599885 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2dc93721-02c2-4553-960f-71955c73b548-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.990709 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" event={"ID":"2dc93721-02c2-4553-960f-71955c73b548","Type":"ContainerDied","Data":"15beaf16eb6d4c8333579c430526cdbcdf8600a933483769a7f7bf413c5c21dc"} Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.991037 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15beaf16eb6d4c8333579c430526cdbcdf8600a933483769a7f7bf413c5c21dc" Dec 06 06:14:15 crc kubenswrapper[4957]: I1206 06:14:15.990787 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-shvb7" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.087741 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5"] Dec 06 06:14:16 crc kubenswrapper[4957]: E1206 06:14:16.088172 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc93721-02c2-4553-960f-71955c73b548" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.088189 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc93721-02c2-4553-960f-71955c73b548" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.088422 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dc93721-02c2-4553-960f-71955c73b548" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.089106 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.095494 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.096277 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.096481 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.096637 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.104706 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5"] Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.108584 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.108677 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.108756 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4gvb\" (UniqueName: \"kubernetes.io/projected/a432bf6f-c43f-4b87-af4b-2af6281f1c25-kube-api-access-n4gvb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.211123 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.211475 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.211637 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4gvb\" (UniqueName: \"kubernetes.io/projected/a432bf6f-c43f-4b87-af4b-2af6281f1c25-kube-api-access-n4gvb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.216403 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.218392 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.230599 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4gvb\" (UniqueName: \"kubernetes.io/projected/a432bf6f-c43f-4b87-af4b-2af6281f1c25-kube-api-access-n4gvb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.435020 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.970356 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5"] Dec 06 06:14:16 crc kubenswrapper[4957]: I1206 06:14:16.977816 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:14:17 crc kubenswrapper[4957]: I1206 06:14:17.001212 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" event={"ID":"a432bf6f-c43f-4b87-af4b-2af6281f1c25","Type":"ContainerStarted","Data":"6a38424ec8930797d72a8f1f3eb0c819b196652fc9cea8e0d73c7f67ff87cf34"} Dec 06 06:14:18 crc kubenswrapper[4957]: I1206 06:14:18.014616 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" event={"ID":"a432bf6f-c43f-4b87-af4b-2af6281f1c25","Type":"ContainerStarted","Data":"55d420dac76c56c551ae077cfa0d6b33fd9c9ed531317ddfa716b6d81a7405a9"} Dec 06 06:14:18 crc kubenswrapper[4957]: I1206 06:14:18.036262 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" podStartSLOduration=1.477651997 podStartE2EDuration="2.036235187s" podCreationTimestamp="2025-12-06 06:14:16 +0000 UTC" firstStartedPulling="2025-12-06 06:14:16.977565009 +0000 UTC m=+2096.627832631" lastFinishedPulling="2025-12-06 06:14:17.536148189 +0000 UTC m=+2097.186415821" observedRunningTime="2025-12-06 06:14:18.029389902 +0000 UTC m=+2097.679657554" watchObservedRunningTime="2025-12-06 06:14:18.036235187 +0000 UTC m=+2097.686502829" Dec 06 06:14:34 crc kubenswrapper[4957]: I1206 06:14:34.048245 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ccb8p"] Dec 06 06:14:34 crc kubenswrapper[4957]: I1206 06:14:34.057997 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ccb8p"] Dec 06 06:14:34 crc kubenswrapper[4957]: I1206 06:14:34.677489 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0cdc965-475d-4d47-b875-9ade77089ae6" path="/var/lib/kubelet/pods/e0cdc965-475d-4d47-b875-9ade77089ae6/volumes" Dec 06 06:14:44 crc kubenswrapper[4957]: I1206 06:14:44.447735 4957 scope.go:117] "RemoveContainer" containerID="5e01d251252448379c8f5523a1847a84733826d8dc28c16afd89780ffaa7c2f7" Dec 06 06:14:44 crc kubenswrapper[4957]: I1206 06:14:44.477454 4957 scope.go:117] "RemoveContainer" containerID="5a401c43cf809772aecd69ff47de3fa392cb5155bbe03671983f6b34538cc667" Dec 06 06:14:44 crc kubenswrapper[4957]: I1206 06:14:44.530959 4957 scope.go:117] "RemoveContainer" containerID="011bcbc7d489f817ef03d3d8001ff4be7679d919b15529df0f612b8d7837199a" Dec 06 06:14:44 crc kubenswrapper[4957]: I1206 06:14:44.582244 4957 scope.go:117] "RemoveContainer" containerID="aa4c3c143e7986c3dcba5a8151acebb153a83551d2983bae7bb477d1b9ae3d7c" Dec 06 06:14:44 crc kubenswrapper[4957]: I1206 06:14:44.620120 4957 scope.go:117] "RemoveContainer" containerID="4f2be6a8216ab58840f826fe9a624980c002648bb8b1059ae92ce567ba4c7cfd" Dec 06 06:14:44 crc kubenswrapper[4957]: I1206 06:14:44.681591 4957 scope.go:117] "RemoveContainer" containerID="f00279cba88970410eea431dcfe35cbbb627a0a43442bdad239de02d570d5396" Dec 06 06:14:44 crc kubenswrapper[4957]: I1206 06:14:44.705264 4957 scope.go:117] "RemoveContainer" containerID="0925acfb1610a9821c6c182618fec73cb46234ef492255c573888698831e3a14" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.170177 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7wvwb"] Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.174075 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.182325 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7wvwb"] Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.267706 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-utilities\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.267884 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-677gr\" (UniqueName: \"kubernetes.io/projected/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-kube-api-access-677gr\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.267947 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-catalog-content\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.369623 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-utilities\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.369749 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-677gr\" (UniqueName: \"kubernetes.io/projected/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-kube-api-access-677gr\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.370119 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-catalog-content\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.370313 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-utilities\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.370351 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-catalog-content\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.390918 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-677gr\" (UniqueName: \"kubernetes.io/projected/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-kube-api-access-677gr\") pod \"community-operators-7wvwb\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:46 crc kubenswrapper[4957]: I1206 06:14:46.538972 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:14:47 crc kubenswrapper[4957]: I1206 06:14:47.093017 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7wvwb"] Dec 06 06:14:47 crc kubenswrapper[4957]: I1206 06:14:47.279460 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvwb" event={"ID":"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7","Type":"ContainerStarted","Data":"d5ee5c360a628d4d83f49d05790fc11c33f07da51db181012d9653cb1e672735"} Dec 06 06:14:48 crc kubenswrapper[4957]: I1206 06:14:48.289283 4957 generic.go:334] "Generic (PLEG): container finished" podID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerID="70de956b182d200aacaa6e4443675040b5e1bf369d4213b6ed774ebf862b2a9a" exitCode=0 Dec 06 06:14:48 crc kubenswrapper[4957]: I1206 06:14:48.289335 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvwb" event={"ID":"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7","Type":"ContainerDied","Data":"70de956b182d200aacaa6e4443675040b5e1bf369d4213b6ed774ebf862b2a9a"} Dec 06 06:14:49 crc kubenswrapper[4957]: I1206 06:14:49.300199 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvwb" event={"ID":"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7","Type":"ContainerStarted","Data":"bf230e240e63149047289847534c86807f9d9ab57b74c4742a52da2dbe8c26cc"} Dec 06 06:14:51 crc kubenswrapper[4957]: I1206 06:14:51.321657 4957 generic.go:334] "Generic (PLEG): container finished" podID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerID="bf230e240e63149047289847534c86807f9d9ab57b74c4742a52da2dbe8c26cc" exitCode=0 Dec 06 06:14:51 crc kubenswrapper[4957]: I1206 06:14:51.321793 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvwb" event={"ID":"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7","Type":"ContainerDied","Data":"bf230e240e63149047289847534c86807f9d9ab57b74c4742a52da2dbe8c26cc"} Dec 06 06:14:55 crc kubenswrapper[4957]: I1206 06:14:55.895051 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jsfx6"] Dec 06 06:14:55 crc kubenswrapper[4957]: I1206 06:14:55.898186 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:55 crc kubenswrapper[4957]: I1206 06:14:55.907515 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jsfx6"] Dec 06 06:14:55 crc kubenswrapper[4957]: I1206 06:14:55.980918 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-utilities\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:55 crc kubenswrapper[4957]: I1206 06:14:55.980989 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcppd\" (UniqueName: \"kubernetes.io/projected/c7597af8-661c-41fa-9ebf-88a44d91f9d6-kube-api-access-gcppd\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:55 crc kubenswrapper[4957]: I1206 06:14:55.981128 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-catalog-content\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.083122 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-catalog-content\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.083253 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-utilities\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.083321 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcppd\" (UniqueName: \"kubernetes.io/projected/c7597af8-661c-41fa-9ebf-88a44d91f9d6-kube-api-access-gcppd\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.083694 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-catalog-content\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.083818 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-utilities\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.105753 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcppd\" (UniqueName: \"kubernetes.io/projected/c7597af8-661c-41fa-9ebf-88a44d91f9d6-kube-api-access-gcppd\") pod \"redhat-marketplace-jsfx6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.226776 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.492166 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7flv6"] Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.494393 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.508485 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7flv6"] Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.592687 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-catalog-content\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.592975 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-utilities\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.593195 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glzxk\" (UniqueName: \"kubernetes.io/projected/c708f364-0665-4bdc-9754-7c2338bf22de-kube-api-access-glzxk\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.694910 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glzxk\" (UniqueName: \"kubernetes.io/projected/c708f364-0665-4bdc-9754-7c2338bf22de-kube-api-access-glzxk\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.695088 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-catalog-content\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.695152 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-utilities\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.695640 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-utilities\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.696261 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-catalog-content\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.725687 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glzxk\" (UniqueName: \"kubernetes.io/projected/c708f364-0665-4bdc-9754-7c2338bf22de-kube-api-access-glzxk\") pod \"redhat-operators-7flv6\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.762528 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jsfx6"] Dec 06 06:14:56 crc kubenswrapper[4957]: W1206 06:14:56.765448 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7597af8_661c_41fa_9ebf_88a44d91f9d6.slice/crio-8c2a2b17bbc92f162c659495bfe57b349c57f4a7ccdb5bd835a23bc9c2576400 WatchSource:0}: Error finding container 8c2a2b17bbc92f162c659495bfe57b349c57f4a7ccdb5bd835a23bc9c2576400: Status 404 returned error can't find the container with id 8c2a2b17bbc92f162c659495bfe57b349c57f4a7ccdb5bd835a23bc9c2576400 Dec 06 06:14:56 crc kubenswrapper[4957]: I1206 06:14:56.825729 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.057451 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-hplhf"] Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.072140 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-hplhf"] Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.351157 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7flv6"] Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.394903 4957 generic.go:334] "Generic (PLEG): container finished" podID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerID="aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093" exitCode=0 Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.395014 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jsfx6" event={"ID":"c7597af8-661c-41fa-9ebf-88a44d91f9d6","Type":"ContainerDied","Data":"aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093"} Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.395127 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jsfx6" event={"ID":"c7597af8-661c-41fa-9ebf-88a44d91f9d6","Type":"ContainerStarted","Data":"8c2a2b17bbc92f162c659495bfe57b349c57f4a7ccdb5bd835a23bc9c2576400"} Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.402521 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvwb" event={"ID":"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7","Type":"ContainerStarted","Data":"5483469de7b47038d63b898a51976fa0c0d26e35d5dc6f5be3b2080ddbe2df80"} Dec 06 06:14:57 crc kubenswrapper[4957]: I1206 06:14:57.406272 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7flv6" event={"ID":"c708f364-0665-4bdc-9754-7c2338bf22de","Type":"ContainerStarted","Data":"f53c6db0fe6d0aad77b347f017ca5fd173cf5400006a84a4b72ba7942a5918bc"} Dec 06 06:14:58 crc kubenswrapper[4957]: I1206 06:14:58.029937 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7wvwb" podStartSLOduration=3.642702737 podStartE2EDuration="12.02990956s" podCreationTimestamp="2025-12-06 06:14:46 +0000 UTC" firstStartedPulling="2025-12-06 06:14:48.291103691 +0000 UTC m=+2127.941371313" lastFinishedPulling="2025-12-06 06:14:56.678310514 +0000 UTC m=+2136.328578136" observedRunningTime="2025-12-06 06:14:57.439916141 +0000 UTC m=+2137.090183793" watchObservedRunningTime="2025-12-06 06:14:58.02990956 +0000 UTC m=+2137.680177192" Dec 06 06:14:58 crc kubenswrapper[4957]: I1206 06:14:58.030614 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xjg58"] Dec 06 06:14:58 crc kubenswrapper[4957]: I1206 06:14:58.042902 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xjg58"] Dec 06 06:14:58 crc kubenswrapper[4957]: I1206 06:14:58.675458 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3" path="/var/lib/kubelet/pods/6db5d8d7-7fc2-437c-a7bb-5c3630eca4f3/volumes" Dec 06 06:14:58 crc kubenswrapper[4957]: I1206 06:14:58.676709 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d31daa9e-4a19-46c6-bdd0-eff90eb42ba8" path="/var/lib/kubelet/pods/d31daa9e-4a19-46c6-bdd0-eff90eb42ba8/volumes" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.149928 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz"] Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.151121 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.154034 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.154122 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.167720 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz"] Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.266709 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bde94ee6-96a0-4433-9b7f-73a34c6085c8-secret-volume\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.267124 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bde94ee6-96a0-4433-9b7f-73a34c6085c8-config-volume\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.267160 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldtg\" (UniqueName: \"kubernetes.io/projected/bde94ee6-96a0-4433-9b7f-73a34c6085c8-kube-api-access-bldtg\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.369079 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bde94ee6-96a0-4433-9b7f-73a34c6085c8-config-volume\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.369209 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldtg\" (UniqueName: \"kubernetes.io/projected/bde94ee6-96a0-4433-9b7f-73a34c6085c8-kube-api-access-bldtg\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.369363 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bde94ee6-96a0-4433-9b7f-73a34c6085c8-secret-volume\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.370005 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bde94ee6-96a0-4433-9b7f-73a34c6085c8-config-volume\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.377041 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bde94ee6-96a0-4433-9b7f-73a34c6085c8-secret-volume\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.387720 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldtg\" (UniqueName: \"kubernetes.io/projected/bde94ee6-96a0-4433-9b7f-73a34c6085c8-kube-api-access-bldtg\") pod \"collect-profiles-29416695-cffcz\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:00 crc kubenswrapper[4957]: I1206 06:15:00.493930 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:02 crc kubenswrapper[4957]: W1206 06:15:02.172053 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbde94ee6_96a0_4433_9b7f_73a34c6085c8.slice/crio-34309d5194fa54c0532f22c3c68faac69a60935f1c14ab4155d86d3d22b43935 WatchSource:0}: Error finding container 34309d5194fa54c0532f22c3c68faac69a60935f1c14ab4155d86d3d22b43935: Status 404 returned error can't find the container with id 34309d5194fa54c0532f22c3c68faac69a60935f1c14ab4155d86d3d22b43935 Dec 06 06:15:02 crc kubenswrapper[4957]: I1206 06:15:02.193136 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz"] Dec 06 06:15:02 crc kubenswrapper[4957]: I1206 06:15:02.460728 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" event={"ID":"bde94ee6-96a0-4433-9b7f-73a34c6085c8","Type":"ContainerStarted","Data":"34309d5194fa54c0532f22c3c68faac69a60935f1c14ab4155d86d3d22b43935"} Dec 06 06:15:05 crc kubenswrapper[4957]: I1206 06:15:05.486984 4957 generic.go:334] "Generic (PLEG): container finished" podID="c708f364-0665-4bdc-9754-7c2338bf22de" containerID="e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b" exitCode=0 Dec 06 06:15:05 crc kubenswrapper[4957]: I1206 06:15:05.487529 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7flv6" event={"ID":"c708f364-0665-4bdc-9754-7c2338bf22de","Type":"ContainerDied","Data":"e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b"} Dec 06 06:15:06 crc kubenswrapper[4957]: I1206 06:15:06.540355 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:15:06 crc kubenswrapper[4957]: I1206 06:15:06.541693 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:15:06 crc kubenswrapper[4957]: I1206 06:15:06.589560 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:15:07 crc kubenswrapper[4957]: I1206 06:15:07.570923 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:15:07 crc kubenswrapper[4957]: I1206 06:15:07.620922 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7wvwb"] Dec 06 06:15:09 crc kubenswrapper[4957]: I1206 06:15:09.526486 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" event={"ID":"bde94ee6-96a0-4433-9b7f-73a34c6085c8","Type":"ContainerStarted","Data":"a761b3c2de714eeb28749d7301343ae955b5c06956633df234b0bf7e07a258e1"} Dec 06 06:15:09 crc kubenswrapper[4957]: I1206 06:15:09.526702 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7wvwb" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="registry-server" containerID="cri-o://5483469de7b47038d63b898a51976fa0c0d26e35d5dc6f5be3b2080ddbe2df80" gracePeriod=2 Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.577707 4957 generic.go:334] "Generic (PLEG): container finished" podID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerID="5483469de7b47038d63b898a51976fa0c0d26e35d5dc6f5be3b2080ddbe2df80" exitCode=0 Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.578454 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvwb" event={"ID":"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7","Type":"ContainerDied","Data":"5483469de7b47038d63b898a51976fa0c0d26e35d5dc6f5be3b2080ddbe2df80"} Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.583490 4957 generic.go:334] "Generic (PLEG): container finished" podID="bde94ee6-96a0-4433-9b7f-73a34c6085c8" containerID="a761b3c2de714eeb28749d7301343ae955b5c06956633df234b0bf7e07a258e1" exitCode=0 Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.583597 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" event={"ID":"bde94ee6-96a0-4433-9b7f-73a34c6085c8","Type":"ContainerDied","Data":"a761b3c2de714eeb28749d7301343ae955b5c06956633df234b0bf7e07a258e1"} Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.780291 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.914685 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-catalog-content\") pod \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.915227 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-677gr\" (UniqueName: \"kubernetes.io/projected/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-kube-api-access-677gr\") pod \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.915328 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-utilities\") pod \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\" (UID: \"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7\") " Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.916001 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-utilities" (OuterVolumeSpecName: "utilities") pod "49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" (UID: "49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.925875 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-kube-api-access-677gr" (OuterVolumeSpecName: "kube-api-access-677gr") pod "49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" (UID: "49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7"). InnerVolumeSpecName "kube-api-access-677gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:13 crc kubenswrapper[4957]: I1206 06:15:13.971918 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" (UID: "49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.017586 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.017634 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-677gr\" (UniqueName: \"kubernetes.io/projected/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-kube-api-access-677gr\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.017649 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.593886 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wvwb" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.593972 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wvwb" event={"ID":"49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7","Type":"ContainerDied","Data":"d5ee5c360a628d4d83f49d05790fc11c33f07da51db181012d9653cb1e672735"} Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.594033 4957 scope.go:117] "RemoveContainer" containerID="5483469de7b47038d63b898a51976fa0c0d26e35d5dc6f5be3b2080ddbe2df80" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.636805 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7wvwb"] Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.637725 4957 scope.go:117] "RemoveContainer" containerID="bf230e240e63149047289847534c86807f9d9ab57b74c4742a52da2dbe8c26cc" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.647222 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7wvwb"] Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.673169 4957 scope.go:117] "RemoveContainer" containerID="70de956b182d200aacaa6e4443675040b5e1bf369d4213b6ed774ebf862b2a9a" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.676781 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" path="/var/lib/kubelet/pods/49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7/volumes" Dec 06 06:15:14 crc kubenswrapper[4957]: I1206 06:15:14.949530 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.139725 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bldtg\" (UniqueName: \"kubernetes.io/projected/bde94ee6-96a0-4433-9b7f-73a34c6085c8-kube-api-access-bldtg\") pod \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.139814 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bde94ee6-96a0-4433-9b7f-73a34c6085c8-secret-volume\") pod \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.140099 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bde94ee6-96a0-4433-9b7f-73a34c6085c8-config-volume\") pod \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\" (UID: \"bde94ee6-96a0-4433-9b7f-73a34c6085c8\") " Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.140889 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bde94ee6-96a0-4433-9b7f-73a34c6085c8-config-volume" (OuterVolumeSpecName: "config-volume") pod "bde94ee6-96a0-4433-9b7f-73a34c6085c8" (UID: "bde94ee6-96a0-4433-9b7f-73a34c6085c8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.141237 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bde94ee6-96a0-4433-9b7f-73a34c6085c8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.149299 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde94ee6-96a0-4433-9b7f-73a34c6085c8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bde94ee6-96a0-4433-9b7f-73a34c6085c8" (UID: "bde94ee6-96a0-4433-9b7f-73a34c6085c8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.149361 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde94ee6-96a0-4433-9b7f-73a34c6085c8-kube-api-access-bldtg" (OuterVolumeSpecName: "kube-api-access-bldtg") pod "bde94ee6-96a0-4433-9b7f-73a34c6085c8" (UID: "bde94ee6-96a0-4433-9b7f-73a34c6085c8"). InnerVolumeSpecName "kube-api-access-bldtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.243307 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bldtg\" (UniqueName: \"kubernetes.io/projected/bde94ee6-96a0-4433-9b7f-73a34c6085c8-kube-api-access-bldtg\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.243343 4957 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bde94ee6-96a0-4433-9b7f-73a34c6085c8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.603780 4957 generic.go:334] "Generic (PLEG): container finished" podID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerID="48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25" exitCode=0 Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.603822 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jsfx6" event={"ID":"c7597af8-661c-41fa-9ebf-88a44d91f9d6","Type":"ContainerDied","Data":"48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25"} Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.608402 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" event={"ID":"bde94ee6-96a0-4433-9b7f-73a34c6085c8","Type":"ContainerDied","Data":"34309d5194fa54c0532f22c3c68faac69a60935f1c14ab4155d86d3d22b43935"} Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.608437 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz" Dec 06 06:15:15 crc kubenswrapper[4957]: I1206 06:15:15.608447 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34309d5194fa54c0532f22c3c68faac69a60935f1c14ab4155d86d3d22b43935" Dec 06 06:15:16 crc kubenswrapper[4957]: I1206 06:15:16.030204 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6"] Dec 06 06:15:16 crc kubenswrapper[4957]: I1206 06:15:16.042547 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416650-9w9j6"] Dec 06 06:15:16 crc kubenswrapper[4957]: I1206 06:15:16.619343 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7flv6" event={"ID":"c708f364-0665-4bdc-9754-7c2338bf22de","Type":"ContainerStarted","Data":"a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb"} Dec 06 06:15:16 crc kubenswrapper[4957]: I1206 06:15:16.673051 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9af915f-4e5b-440d-a3b3-534d16b1f686" path="/var/lib/kubelet/pods/f9af915f-4e5b-440d-a3b3-534d16b1f686/volumes" Dec 06 06:15:18 crc kubenswrapper[4957]: I1206 06:15:18.639620 4957 generic.go:334] "Generic (PLEG): container finished" podID="c708f364-0665-4bdc-9754-7c2338bf22de" containerID="a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb" exitCode=0 Dec 06 06:15:18 crc kubenswrapper[4957]: I1206 06:15:18.639673 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7flv6" event={"ID":"c708f364-0665-4bdc-9754-7c2338bf22de","Type":"ContainerDied","Data":"a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb"} Dec 06 06:15:20 crc kubenswrapper[4957]: I1206 06:15:20.699099 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jsfx6" event={"ID":"c7597af8-661c-41fa-9ebf-88a44d91f9d6","Type":"ContainerStarted","Data":"9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0"} Dec 06 06:15:21 crc kubenswrapper[4957]: I1206 06:15:21.711904 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7flv6" event={"ID":"c708f364-0665-4bdc-9754-7c2338bf22de","Type":"ContainerStarted","Data":"4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0"} Dec 06 06:15:21 crc kubenswrapper[4957]: I1206 06:15:21.738268 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jsfx6" podStartSLOduration=4.163145315 podStartE2EDuration="26.738247323s" podCreationTimestamp="2025-12-06 06:14:55 +0000 UTC" firstStartedPulling="2025-12-06 06:14:57.40654997 +0000 UTC m=+2137.056817612" lastFinishedPulling="2025-12-06 06:15:19.981651988 +0000 UTC m=+2159.631919620" observedRunningTime="2025-12-06 06:15:20.720178472 +0000 UTC m=+2160.370446104" watchObservedRunningTime="2025-12-06 06:15:21.738247323 +0000 UTC m=+2161.388514955" Dec 06 06:15:21 crc kubenswrapper[4957]: I1206 06:15:21.740245 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7flv6" podStartSLOduration=18.995472751 podStartE2EDuration="25.740235316s" podCreationTimestamp="2025-12-06 06:14:56 +0000 UTC" firstStartedPulling="2025-12-06 06:15:13.699779816 +0000 UTC m=+2153.350047448" lastFinishedPulling="2025-12-06 06:15:20.444542381 +0000 UTC m=+2160.094810013" observedRunningTime="2025-12-06 06:15:21.730201855 +0000 UTC m=+2161.380469497" watchObservedRunningTime="2025-12-06 06:15:21.740235316 +0000 UTC m=+2161.390502948" Dec 06 06:15:26 crc kubenswrapper[4957]: I1206 06:15:26.227587 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:15:26 crc kubenswrapper[4957]: I1206 06:15:26.228122 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:15:26 crc kubenswrapper[4957]: I1206 06:15:26.270807 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:15:26 crc kubenswrapper[4957]: I1206 06:15:26.805311 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:15:26 crc kubenswrapper[4957]: I1206 06:15:26.826659 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:15:26 crc kubenswrapper[4957]: I1206 06:15:26.828017 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:15:27 crc kubenswrapper[4957]: I1206 06:15:27.874367 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7flv6" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="registry-server" probeResult="failure" output=< Dec 06 06:15:27 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 06:15:27 crc kubenswrapper[4957]: > Dec 06 06:15:28 crc kubenswrapper[4957]: I1206 06:15:28.417122 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jsfx6"] Dec 06 06:15:28 crc kubenswrapper[4957]: I1206 06:15:28.772658 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jsfx6" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="registry-server" containerID="cri-o://9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0" gracePeriod=2 Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.720936 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.738986 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-utilities\") pod \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.739059 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcppd\" (UniqueName: \"kubernetes.io/projected/c7597af8-661c-41fa-9ebf-88a44d91f9d6-kube-api-access-gcppd\") pod \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.739087 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-catalog-content\") pod \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\" (UID: \"c7597af8-661c-41fa-9ebf-88a44d91f9d6\") " Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.739975 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-utilities" (OuterVolumeSpecName: "utilities") pod "c7597af8-661c-41fa-9ebf-88a44d91f9d6" (UID: "c7597af8-661c-41fa-9ebf-88a44d91f9d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.760787 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7597af8-661c-41fa-9ebf-88a44d91f9d6" (UID: "c7597af8-661c-41fa-9ebf-88a44d91f9d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.764223 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7597af8-661c-41fa-9ebf-88a44d91f9d6-kube-api-access-gcppd" (OuterVolumeSpecName: "kube-api-access-gcppd") pod "c7597af8-661c-41fa-9ebf-88a44d91f9d6" (UID: "c7597af8-661c-41fa-9ebf-88a44d91f9d6"). InnerVolumeSpecName "kube-api-access-gcppd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.781743 4957 generic.go:334] "Generic (PLEG): container finished" podID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerID="9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0" exitCode=0 Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.781781 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jsfx6" event={"ID":"c7597af8-661c-41fa-9ebf-88a44d91f9d6","Type":"ContainerDied","Data":"9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0"} Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.781806 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jsfx6" event={"ID":"c7597af8-661c-41fa-9ebf-88a44d91f9d6","Type":"ContainerDied","Data":"8c2a2b17bbc92f162c659495bfe57b349c57f4a7ccdb5bd835a23bc9c2576400"} Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.781822 4957 scope.go:117] "RemoveContainer" containerID="9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.781961 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jsfx6" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.832374 4957 scope.go:117] "RemoveContainer" containerID="48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.841132 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.841163 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcppd\" (UniqueName: \"kubernetes.io/projected/c7597af8-661c-41fa-9ebf-88a44d91f9d6-kube-api-access-gcppd\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.841172 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7597af8-661c-41fa-9ebf-88a44d91f9d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.849095 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jsfx6"] Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.867877 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jsfx6"] Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.869135 4957 scope.go:117] "RemoveContainer" containerID="aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.894467 4957 scope.go:117] "RemoveContainer" containerID="9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0" Dec 06 06:15:29 crc kubenswrapper[4957]: E1206 06:15:29.894954 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0\": container with ID starting with 9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0 not found: ID does not exist" containerID="9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.894985 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0"} err="failed to get container status \"9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0\": rpc error: code = NotFound desc = could not find container \"9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0\": container with ID starting with 9fdb2d3dfcb5b231e8d2a596246e3a161ba15a5f4a0d97505cc6593a2bc7cff0 not found: ID does not exist" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.895010 4957 scope.go:117] "RemoveContainer" containerID="48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25" Dec 06 06:15:29 crc kubenswrapper[4957]: E1206 06:15:29.895278 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25\": container with ID starting with 48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25 not found: ID does not exist" containerID="48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.895300 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25"} err="failed to get container status \"48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25\": rpc error: code = NotFound desc = could not find container \"48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25\": container with ID starting with 48237474e7a3716fc7ec4f50742aa853343adf398f10ec0f52fcfb4aa6fcec25 not found: ID does not exist" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.895316 4957 scope.go:117] "RemoveContainer" containerID="aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093" Dec 06 06:15:29 crc kubenswrapper[4957]: E1206 06:15:29.895670 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093\": container with ID starting with aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093 not found: ID does not exist" containerID="aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093" Dec 06 06:15:29 crc kubenswrapper[4957]: I1206 06:15:29.895687 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093"} err="failed to get container status \"aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093\": rpc error: code = NotFound desc = could not find container \"aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093\": container with ID starting with aa2cf246e258673c2b67aa1a045679010ff461dfb230e3ee02f3298df3ae5093 not found: ID does not exist" Dec 06 06:15:30 crc kubenswrapper[4957]: I1206 06:15:30.673071 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" path="/var/lib/kubelet/pods/c7597af8-661c-41fa-9ebf-88a44d91f9d6/volumes" Dec 06 06:15:36 crc kubenswrapper[4957]: I1206 06:15:36.879059 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:15:36 crc kubenswrapper[4957]: I1206 06:15:36.930595 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:15:37 crc kubenswrapper[4957]: I1206 06:15:37.111016 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7flv6"] Dec 06 06:15:38 crc kubenswrapper[4957]: I1206 06:15:38.868788 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7flv6" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="registry-server" containerID="cri-o://4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0" gracePeriod=2 Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.292405 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.412344 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-utilities\") pod \"c708f364-0665-4bdc-9754-7c2338bf22de\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.412989 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-catalog-content\") pod \"c708f364-0665-4bdc-9754-7c2338bf22de\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.413212 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glzxk\" (UniqueName: \"kubernetes.io/projected/c708f364-0665-4bdc-9754-7c2338bf22de-kube-api-access-glzxk\") pod \"c708f364-0665-4bdc-9754-7c2338bf22de\" (UID: \"c708f364-0665-4bdc-9754-7c2338bf22de\") " Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.413711 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-utilities" (OuterVolumeSpecName: "utilities") pod "c708f364-0665-4bdc-9754-7c2338bf22de" (UID: "c708f364-0665-4bdc-9754-7c2338bf22de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.419322 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c708f364-0665-4bdc-9754-7c2338bf22de-kube-api-access-glzxk" (OuterVolumeSpecName: "kube-api-access-glzxk") pod "c708f364-0665-4bdc-9754-7c2338bf22de" (UID: "c708f364-0665-4bdc-9754-7c2338bf22de"). InnerVolumeSpecName "kube-api-access-glzxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.515206 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glzxk\" (UniqueName: \"kubernetes.io/projected/c708f364-0665-4bdc-9754-7c2338bf22de-kube-api-access-glzxk\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.515261 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.533784 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c708f364-0665-4bdc-9754-7c2338bf22de" (UID: "c708f364-0665-4bdc-9754-7c2338bf22de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.617646 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c708f364-0665-4bdc-9754-7c2338bf22de-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.883301 4957 generic.go:334] "Generic (PLEG): container finished" podID="c708f364-0665-4bdc-9754-7c2338bf22de" containerID="4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0" exitCode=0 Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.883381 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7flv6" event={"ID":"c708f364-0665-4bdc-9754-7c2338bf22de","Type":"ContainerDied","Data":"4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0"} Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.883416 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7flv6" event={"ID":"c708f364-0665-4bdc-9754-7c2338bf22de","Type":"ContainerDied","Data":"f53c6db0fe6d0aad77b347f017ca5fd173cf5400006a84a4b72ba7942a5918bc"} Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.883439 4957 scope.go:117] "RemoveContainer" containerID="4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.883679 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7flv6" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.915921 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7flv6"] Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.923521 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7flv6"] Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.933051 4957 scope.go:117] "RemoveContainer" containerID="a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb" Dec 06 06:15:39 crc kubenswrapper[4957]: I1206 06:15:39.960978 4957 scope.go:117] "RemoveContainer" containerID="e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b" Dec 06 06:15:40 crc kubenswrapper[4957]: I1206 06:15:40.001666 4957 scope.go:117] "RemoveContainer" containerID="4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0" Dec 06 06:15:40 crc kubenswrapper[4957]: E1206 06:15:40.002360 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0\": container with ID starting with 4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0 not found: ID does not exist" containerID="4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0" Dec 06 06:15:40 crc kubenswrapper[4957]: I1206 06:15:40.002405 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0"} err="failed to get container status \"4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0\": rpc error: code = NotFound desc = could not find container \"4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0\": container with ID starting with 4c88c213e7703fab9f082ce07c5ff5c670161e61bed47b11e19b216c792437d0 not found: ID does not exist" Dec 06 06:15:40 crc kubenswrapper[4957]: I1206 06:15:40.002429 4957 scope.go:117] "RemoveContainer" containerID="a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb" Dec 06 06:15:40 crc kubenswrapper[4957]: E1206 06:15:40.004635 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb\": container with ID starting with a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb not found: ID does not exist" containerID="a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb" Dec 06 06:15:40 crc kubenswrapper[4957]: I1206 06:15:40.004769 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb"} err="failed to get container status \"a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb\": rpc error: code = NotFound desc = could not find container \"a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb\": container with ID starting with a3ecc73c7b1a7befd75ab1757c7650a65b3b295ff074c8f710986f19e39b60eb not found: ID does not exist" Dec 06 06:15:40 crc kubenswrapper[4957]: I1206 06:15:40.004823 4957 scope.go:117] "RemoveContainer" containerID="e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b" Dec 06 06:15:40 crc kubenswrapper[4957]: E1206 06:15:40.005472 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b\": container with ID starting with e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b not found: ID does not exist" containerID="e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b" Dec 06 06:15:40 crc kubenswrapper[4957]: I1206 06:15:40.005558 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b"} err="failed to get container status \"e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b\": rpc error: code = NotFound desc = could not find container \"e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b\": container with ID starting with e2ae6e70d029eee35c0be33cad2be5fefc795ff9a715a1634bd9eb797a538d3b not found: ID does not exist" Dec 06 06:15:40 crc kubenswrapper[4957]: I1206 06:15:40.691466 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" path="/var/lib/kubelet/pods/c708f364-0665-4bdc-9754-7c2338bf22de/volumes" Dec 06 06:15:42 crc kubenswrapper[4957]: I1206 06:15:42.046242 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-56t98"] Dec 06 06:15:42 crc kubenswrapper[4957]: I1206 06:15:42.057140 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-56t98"] Dec 06 06:15:42 crc kubenswrapper[4957]: I1206 06:15:42.673646 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c5dbab8-a215-48b4-beea-7a5ae2715156" path="/var/lib/kubelet/pods/9c5dbab8-a215-48b4-beea-7a5ae2715156/volumes" Dec 06 06:15:44 crc kubenswrapper[4957]: I1206 06:15:44.840472 4957 scope.go:117] "RemoveContainer" containerID="d0a9a72d7eef5d65a50afd7b21162ef1e895c03133cf87e5d26afa27076760a4" Dec 06 06:15:44 crc kubenswrapper[4957]: I1206 06:15:44.889730 4957 scope.go:117] "RemoveContainer" containerID="e6b39df616f7b247d0534fa5dbdc7b2cc9c3de3a6b32187da0775b5c1e0a0c72" Dec 06 06:15:44 crc kubenswrapper[4957]: I1206 06:15:44.949883 4957 scope.go:117] "RemoveContainer" containerID="ae82e43bea073eaff0c99228d7792633beab7e0865f7a8b12ba11fd1cd1f1c62" Dec 06 06:15:45 crc kubenswrapper[4957]: I1206 06:15:45.009414 4957 scope.go:117] "RemoveContainer" containerID="cef6bc9892b207b2f3b961a1b79ffebf77071515014aeeb07dbe38b4fc6a239c" Dec 06 06:16:01 crc kubenswrapper[4957]: I1206 06:16:01.077586 4957 generic.go:334] "Generic (PLEG): container finished" podID="a432bf6f-c43f-4b87-af4b-2af6281f1c25" containerID="55d420dac76c56c551ae077cfa0d6b33fd9c9ed531317ddfa716b6d81a7405a9" exitCode=0 Dec 06 06:16:01 crc kubenswrapper[4957]: I1206 06:16:01.077672 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" event={"ID":"a432bf6f-c43f-4b87-af4b-2af6281f1c25","Type":"ContainerDied","Data":"55d420dac76c56c551ae077cfa0d6b33fd9c9ed531317ddfa716b6d81a7405a9"} Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.611343 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.685361 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-inventory\") pod \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.685566 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-ssh-key\") pod \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.685652 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4gvb\" (UniqueName: \"kubernetes.io/projected/a432bf6f-c43f-4b87-af4b-2af6281f1c25-kube-api-access-n4gvb\") pod \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\" (UID: \"a432bf6f-c43f-4b87-af4b-2af6281f1c25\") " Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.691943 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a432bf6f-c43f-4b87-af4b-2af6281f1c25-kube-api-access-n4gvb" (OuterVolumeSpecName: "kube-api-access-n4gvb") pod "a432bf6f-c43f-4b87-af4b-2af6281f1c25" (UID: "a432bf6f-c43f-4b87-af4b-2af6281f1c25"). InnerVolumeSpecName "kube-api-access-n4gvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.714117 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a432bf6f-c43f-4b87-af4b-2af6281f1c25" (UID: "a432bf6f-c43f-4b87-af4b-2af6281f1c25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.714810 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-inventory" (OuterVolumeSpecName: "inventory") pod "a432bf6f-c43f-4b87-af4b-2af6281f1c25" (UID: "a432bf6f-c43f-4b87-af4b-2af6281f1c25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.788748 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.788808 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4gvb\" (UniqueName: \"kubernetes.io/projected/a432bf6f-c43f-4b87-af4b-2af6281f1c25-kube-api-access-n4gvb\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:02 crc kubenswrapper[4957]: I1206 06:16:02.788856 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a432bf6f-c43f-4b87-af4b-2af6281f1c25-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.103225 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" event={"ID":"a432bf6f-c43f-4b87-af4b-2af6281f1c25","Type":"ContainerDied","Data":"6a38424ec8930797d72a8f1f3eb0c819b196652fc9cea8e0d73c7f67ff87cf34"} Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.103266 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a38424ec8930797d72a8f1f3eb0c819b196652fc9cea8e0d73c7f67ff87cf34" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.103276 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196220 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z"] Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196730 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="extract-content" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196755 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="extract-content" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196797 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a432bf6f-c43f-4b87-af4b-2af6281f1c25" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196810 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a432bf6f-c43f-4b87-af4b-2af6281f1c25" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196824 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196853 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196873 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="extract-utilities" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196883 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="extract-utilities" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196894 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196905 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196928 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde94ee6-96a0-4433-9b7f-73a34c6085c8" containerName="collect-profiles" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196941 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde94ee6-96a0-4433-9b7f-73a34c6085c8" containerName="collect-profiles" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196957 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196965 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.196980 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="extract-utilities" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.196988 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="extract-utilities" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.197003 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="extract-content" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197011 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="extract-content" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.197026 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="extract-utilities" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197034 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="extract-utilities" Dec 06 06:16:03 crc kubenswrapper[4957]: E1206 06:16:03.197046 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="extract-content" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197054 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="extract-content" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197329 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde94ee6-96a0-4433-9b7f-73a34c6085c8" containerName="collect-profiles" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197344 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a432bf6f-c43f-4b87-af4b-2af6281f1c25" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197354 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c708f364-0665-4bdc-9754-7c2338bf22de" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197376 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7597af8-661c-41fa-9ebf-88a44d91f9d6" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.197402 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b3fc8f-85f7-44c7-9eee-f5ff7f535ef7" containerName="registry-server" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.198242 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.202862 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.203484 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.204189 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.209945 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.212427 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z"] Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.299896 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhbp5\" (UniqueName: \"kubernetes.io/projected/7f74c7c3-a665-4958-aaad-15ed4e962fd0-kube-api-access-xhbp5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.299967 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.300058 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.402402 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhbp5\" (UniqueName: \"kubernetes.io/projected/7f74c7c3-a665-4958-aaad-15ed4e962fd0-kube-api-access-xhbp5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.402517 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.402626 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.407138 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.408520 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.424137 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhbp5\" (UniqueName: \"kubernetes.io/projected/7f74c7c3-a665-4958-aaad-15ed4e962fd0-kube-api-access-xhbp5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f799z\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:03 crc kubenswrapper[4957]: I1206 06:16:03.522659 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:04 crc kubenswrapper[4957]: I1206 06:16:04.053828 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z"] Dec 06 06:16:04 crc kubenswrapper[4957]: I1206 06:16:04.114882 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" event={"ID":"7f74c7c3-a665-4958-aaad-15ed4e962fd0","Type":"ContainerStarted","Data":"d58b17bbdbc0b91620eefed10c710a96ac4acaf56e191eb986649c4a498e014f"} Dec 06 06:16:06 crc kubenswrapper[4957]: I1206 06:16:06.133340 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" event={"ID":"7f74c7c3-a665-4958-aaad-15ed4e962fd0","Type":"ContainerStarted","Data":"9b4a8b4a8ff0fb89b4c64093fe4ca5c3a918a6a1916b76269965ce021ddb4645"} Dec 06 06:16:06 crc kubenswrapper[4957]: I1206 06:16:06.151162 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" podStartSLOduration=2.332428538 podStartE2EDuration="3.151140673s" podCreationTimestamp="2025-12-06 06:16:03 +0000 UTC" firstStartedPulling="2025-12-06 06:16:04.069883486 +0000 UTC m=+2203.720151118" lastFinishedPulling="2025-12-06 06:16:04.888595611 +0000 UTC m=+2204.538863253" observedRunningTime="2025-12-06 06:16:06.148649517 +0000 UTC m=+2205.798917159" watchObservedRunningTime="2025-12-06 06:16:06.151140673 +0000 UTC m=+2205.801408305" Dec 06 06:16:10 crc kubenswrapper[4957]: I1206 06:16:10.178760 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" event={"ID":"7f74c7c3-a665-4958-aaad-15ed4e962fd0","Type":"ContainerDied","Data":"9b4a8b4a8ff0fb89b4c64093fe4ca5c3a918a6a1916b76269965ce021ddb4645"} Dec 06 06:16:10 crc kubenswrapper[4957]: I1206 06:16:10.180022 4957 generic.go:334] "Generic (PLEG): container finished" podID="7f74c7c3-a665-4958-aaad-15ed4e962fd0" containerID="9b4a8b4a8ff0fb89b4c64093fe4ca5c3a918a6a1916b76269965ce021ddb4645" exitCode=0 Dec 06 06:16:10 crc kubenswrapper[4957]: I1206 06:16:10.180929 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:16:10 crc kubenswrapper[4957]: I1206 06:16:10.181314 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.672540 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.868290 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-ssh-key\") pod \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.868503 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhbp5\" (UniqueName: \"kubernetes.io/projected/7f74c7c3-a665-4958-aaad-15ed4e962fd0-kube-api-access-xhbp5\") pod \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.868706 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-inventory\") pod \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\" (UID: \"7f74c7c3-a665-4958-aaad-15ed4e962fd0\") " Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.877746 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f74c7c3-a665-4958-aaad-15ed4e962fd0-kube-api-access-xhbp5" (OuterVolumeSpecName: "kube-api-access-xhbp5") pod "7f74c7c3-a665-4958-aaad-15ed4e962fd0" (UID: "7f74c7c3-a665-4958-aaad-15ed4e962fd0"). InnerVolumeSpecName "kube-api-access-xhbp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.900777 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-inventory" (OuterVolumeSpecName: "inventory") pod "7f74c7c3-a665-4958-aaad-15ed4e962fd0" (UID: "7f74c7c3-a665-4958-aaad-15ed4e962fd0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.901106 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f74c7c3-a665-4958-aaad-15ed4e962fd0" (UID: "7f74c7c3-a665-4958-aaad-15ed4e962fd0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.971791 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhbp5\" (UniqueName: \"kubernetes.io/projected/7f74c7c3-a665-4958-aaad-15ed4e962fd0-kube-api-access-xhbp5\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.971854 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:11 crc kubenswrapper[4957]: I1206 06:16:11.971869 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f74c7c3-a665-4958-aaad-15ed4e962fd0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.201942 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" event={"ID":"7f74c7c3-a665-4958-aaad-15ed4e962fd0","Type":"ContainerDied","Data":"d58b17bbdbc0b91620eefed10c710a96ac4acaf56e191eb986649c4a498e014f"} Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.201990 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d58b17bbdbc0b91620eefed10c710a96ac4acaf56e191eb986649c4a498e014f" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.202003 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f799z" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.287439 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487"] Dec 06 06:16:12 crc kubenswrapper[4957]: E1206 06:16:12.287993 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f74c7c3-a665-4958-aaad-15ed4e962fd0" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.288013 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f74c7c3-a665-4958-aaad-15ed4e962fd0" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.288314 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f74c7c3-a665-4958-aaad-15ed4e962fd0" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.289111 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.291451 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.291775 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.292098 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.292373 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.297106 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487"] Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.379110 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.379338 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfp7q\" (UniqueName: \"kubernetes.io/projected/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-kube-api-access-jfp7q\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.379433 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.480911 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfp7q\" (UniqueName: \"kubernetes.io/projected/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-kube-api-access-jfp7q\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.480990 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.481060 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.485321 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.487795 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.499909 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfp7q\" (UniqueName: \"kubernetes.io/projected/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-kube-api-access-jfp7q\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-2n487\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:12 crc kubenswrapper[4957]: I1206 06:16:12.607246 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:16:13 crc kubenswrapper[4957]: I1206 06:16:13.143055 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487"] Dec 06 06:16:13 crc kubenswrapper[4957]: W1206 06:16:13.148981 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87515d38_f8c8_4dd5_9373_fd2bd9ef7992.slice/crio-78c165aed7614bbacad8905fa248c228f2867b2af5142b33fb4ac5aefbf87c1f WatchSource:0}: Error finding container 78c165aed7614bbacad8905fa248c228f2867b2af5142b33fb4ac5aefbf87c1f: Status 404 returned error can't find the container with id 78c165aed7614bbacad8905fa248c228f2867b2af5142b33fb4ac5aefbf87c1f Dec 06 06:16:13 crc kubenswrapper[4957]: I1206 06:16:13.211865 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" event={"ID":"87515d38-f8c8-4dd5-9373-fd2bd9ef7992","Type":"ContainerStarted","Data":"78c165aed7614bbacad8905fa248c228f2867b2af5142b33fb4ac5aefbf87c1f"} Dec 06 06:16:16 crc kubenswrapper[4957]: I1206 06:16:16.242923 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" event={"ID":"87515d38-f8c8-4dd5-9373-fd2bd9ef7992","Type":"ContainerStarted","Data":"f2f224304fffb02c8c19012238ae0f4f52870975ac04f33ad27578d8ffd58ff4"} Dec 06 06:16:16 crc kubenswrapper[4957]: I1206 06:16:16.263605 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" podStartSLOduration=2.144672404 podStartE2EDuration="4.263585953s" podCreationTimestamp="2025-12-06 06:16:12 +0000 UTC" firstStartedPulling="2025-12-06 06:16:13.152567613 +0000 UTC m=+2212.802835245" lastFinishedPulling="2025-12-06 06:16:15.271481162 +0000 UTC m=+2214.921748794" observedRunningTime="2025-12-06 06:16:16.263383338 +0000 UTC m=+2215.913650980" watchObservedRunningTime="2025-12-06 06:16:16.263585953 +0000 UTC m=+2215.913853595" Dec 06 06:16:40 crc kubenswrapper[4957]: I1206 06:16:40.181339 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:16:40 crc kubenswrapper[4957]: I1206 06:16:40.182024 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:16:58 crc kubenswrapper[4957]: I1206 06:16:58.693660 4957 generic.go:334] "Generic (PLEG): container finished" podID="87515d38-f8c8-4dd5-9373-fd2bd9ef7992" containerID="f2f224304fffb02c8c19012238ae0f4f52870975ac04f33ad27578d8ffd58ff4" exitCode=0 Dec 06 06:16:58 crc kubenswrapper[4957]: I1206 06:16:58.693781 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" event={"ID":"87515d38-f8c8-4dd5-9373-fd2bd9ef7992","Type":"ContainerDied","Data":"f2f224304fffb02c8c19012238ae0f4f52870975ac04f33ad27578d8ffd58ff4"} Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.247907 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.416897 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-inventory\") pod \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.416997 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfp7q\" (UniqueName: \"kubernetes.io/projected/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-kube-api-access-jfp7q\") pod \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.417205 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-ssh-key\") pod \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\" (UID: \"87515d38-f8c8-4dd5-9373-fd2bd9ef7992\") " Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.427825 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-kube-api-access-jfp7q" (OuterVolumeSpecName: "kube-api-access-jfp7q") pod "87515d38-f8c8-4dd5-9373-fd2bd9ef7992" (UID: "87515d38-f8c8-4dd5-9373-fd2bd9ef7992"). InnerVolumeSpecName "kube-api-access-jfp7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.449450 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87515d38-f8c8-4dd5-9373-fd2bd9ef7992" (UID: "87515d38-f8c8-4dd5-9373-fd2bd9ef7992"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.449976 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-inventory" (OuterVolumeSpecName: "inventory") pod "87515d38-f8c8-4dd5-9373-fd2bd9ef7992" (UID: "87515d38-f8c8-4dd5-9373-fd2bd9ef7992"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.520378 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.520408 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.520418 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfp7q\" (UniqueName: \"kubernetes.io/projected/87515d38-f8c8-4dd5-9373-fd2bd9ef7992-kube-api-access-jfp7q\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.722667 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" event={"ID":"87515d38-f8c8-4dd5-9373-fd2bd9ef7992","Type":"ContainerDied","Data":"78c165aed7614bbacad8905fa248c228f2867b2af5142b33fb4ac5aefbf87c1f"} Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.723069 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78c165aed7614bbacad8905fa248c228f2867b2af5142b33fb4ac5aefbf87c1f" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.722734 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-2n487" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.941146 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc"] Dec 06 06:17:00 crc kubenswrapper[4957]: E1206 06:17:00.941942 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87515d38-f8c8-4dd5-9373-fd2bd9ef7992" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.941989 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="87515d38-f8c8-4dd5-9373-fd2bd9ef7992" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.942365 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="87515d38-f8c8-4dd5-9373-fd2bd9ef7992" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.943639 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.947761 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.947947 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.948235 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.948438 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:17:00 crc kubenswrapper[4957]: I1206 06:17:00.950425 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc"] Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.031428 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.031529 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhvpg\" (UniqueName: \"kubernetes.io/projected/2cb8adc2-7a82-4969-93d3-abac49e01d24-kube-api-access-rhvpg\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.031905 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.132707 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.132780 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhvpg\" (UniqueName: \"kubernetes.io/projected/2cb8adc2-7a82-4969-93d3-abac49e01d24-kube-api-access-rhvpg\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.132899 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.136866 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.138532 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.153596 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhvpg\" (UniqueName: \"kubernetes.io/projected/2cb8adc2-7a82-4969-93d3-abac49e01d24-kube-api-access-rhvpg\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.308788 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:01 crc kubenswrapper[4957]: I1206 06:17:01.854207 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc"] Dec 06 06:17:02 crc kubenswrapper[4957]: I1206 06:17:02.748475 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" event={"ID":"2cb8adc2-7a82-4969-93d3-abac49e01d24","Type":"ContainerStarted","Data":"c622c93b256620d403dca3ca20f15f7efa72f5957de183414cfa119be9cfb0b5"} Dec 06 06:17:03 crc kubenswrapper[4957]: I1206 06:17:03.759623 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" event={"ID":"2cb8adc2-7a82-4969-93d3-abac49e01d24","Type":"ContainerStarted","Data":"7bf7e05bf046be7f2bc8859dce8cfd19cc8f961ae385426b571f02a7a40639df"} Dec 06 06:17:03 crc kubenswrapper[4957]: I1206 06:17:03.782820 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" podStartSLOduration=2.358182709 podStartE2EDuration="3.782802834s" podCreationTimestamp="2025-12-06 06:17:00 +0000 UTC" firstStartedPulling="2025-12-06 06:17:01.86406917 +0000 UTC m=+2261.514336812" lastFinishedPulling="2025-12-06 06:17:03.288689285 +0000 UTC m=+2262.938956937" observedRunningTime="2025-12-06 06:17:03.776296899 +0000 UTC m=+2263.426564591" watchObservedRunningTime="2025-12-06 06:17:03.782802834 +0000 UTC m=+2263.433070466" Dec 06 06:17:10 crc kubenswrapper[4957]: I1206 06:17:10.180747 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:17:10 crc kubenswrapper[4957]: I1206 06:17:10.181477 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:17:10 crc kubenswrapper[4957]: I1206 06:17:10.181535 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:17:10 crc kubenswrapper[4957]: I1206 06:17:10.182434 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:17:10 crc kubenswrapper[4957]: I1206 06:17:10.182491 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" gracePeriod=600 Dec 06 06:17:11 crc kubenswrapper[4957]: E1206 06:17:11.521871 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:17:11 crc kubenswrapper[4957]: I1206 06:17:11.837290 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" exitCode=0 Dec 06 06:17:11 crc kubenswrapper[4957]: I1206 06:17:11.837328 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092"} Dec 06 06:17:11 crc kubenswrapper[4957]: I1206 06:17:11.837358 4957 scope.go:117] "RemoveContainer" containerID="5117ce9e55733277f221b30cc29bce4ae29d818634565ff5fe20d5464771feb5" Dec 06 06:17:11 crc kubenswrapper[4957]: I1206 06:17:11.837986 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:17:11 crc kubenswrapper[4957]: E1206 06:17:11.838303 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:17:25 crc kubenswrapper[4957]: I1206 06:17:25.662704 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:17:25 crc kubenswrapper[4957]: E1206 06:17:25.663594 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:17:39 crc kubenswrapper[4957]: I1206 06:17:39.662546 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:17:39 crc kubenswrapper[4957]: E1206 06:17:39.663412 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:17:51 crc kubenswrapper[4957]: I1206 06:17:51.662763 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:17:51 crc kubenswrapper[4957]: E1206 06:17:51.663513 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:17:54 crc kubenswrapper[4957]: I1206 06:17:54.209609 4957 generic.go:334] "Generic (PLEG): container finished" podID="2cb8adc2-7a82-4969-93d3-abac49e01d24" containerID="7bf7e05bf046be7f2bc8859dce8cfd19cc8f961ae385426b571f02a7a40639df" exitCode=0 Dec 06 06:17:54 crc kubenswrapper[4957]: I1206 06:17:54.209727 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" event={"ID":"2cb8adc2-7a82-4969-93d3-abac49e01d24","Type":"ContainerDied","Data":"7bf7e05bf046be7f2bc8859dce8cfd19cc8f961ae385426b571f02a7a40639df"} Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.647127 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.810594 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhvpg\" (UniqueName: \"kubernetes.io/projected/2cb8adc2-7a82-4969-93d3-abac49e01d24-kube-api-access-rhvpg\") pod \"2cb8adc2-7a82-4969-93d3-abac49e01d24\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.810653 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-inventory\") pod \"2cb8adc2-7a82-4969-93d3-abac49e01d24\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.810744 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-ssh-key\") pod \"2cb8adc2-7a82-4969-93d3-abac49e01d24\" (UID: \"2cb8adc2-7a82-4969-93d3-abac49e01d24\") " Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.829517 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb8adc2-7a82-4969-93d3-abac49e01d24-kube-api-access-rhvpg" (OuterVolumeSpecName: "kube-api-access-rhvpg") pod "2cb8adc2-7a82-4969-93d3-abac49e01d24" (UID: "2cb8adc2-7a82-4969-93d3-abac49e01d24"). InnerVolumeSpecName "kube-api-access-rhvpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.847250 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-inventory" (OuterVolumeSpecName: "inventory") pod "2cb8adc2-7a82-4969-93d3-abac49e01d24" (UID: "2cb8adc2-7a82-4969-93d3-abac49e01d24"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.857573 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2cb8adc2-7a82-4969-93d3-abac49e01d24" (UID: "2cb8adc2-7a82-4969-93d3-abac49e01d24"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.912621 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.912651 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb8adc2-7a82-4969-93d3-abac49e01d24-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:55 crc kubenswrapper[4957]: I1206 06:17:55.912661 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhvpg\" (UniqueName: \"kubernetes.io/projected/2cb8adc2-7a82-4969-93d3-abac49e01d24-kube-api-access-rhvpg\") on node \"crc\" DevicePath \"\"" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.233111 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" event={"ID":"2cb8adc2-7a82-4969-93d3-abac49e01d24","Type":"ContainerDied","Data":"c622c93b256620d403dca3ca20f15f7efa72f5957de183414cfa119be9cfb0b5"} Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.233521 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c622c93b256620d403dca3ca20f15f7efa72f5957de183414cfa119be9cfb0b5" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.233225 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.337588 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c7hml"] Dec 06 06:17:56 crc kubenswrapper[4957]: E1206 06:17:56.340008 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb8adc2-7a82-4969-93d3-abac49e01d24" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.340042 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb8adc2-7a82-4969-93d3-abac49e01d24" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.340271 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb8adc2-7a82-4969-93d3-abac49e01d24" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.341158 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.343362 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.344424 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.344628 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.345356 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.347108 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c7hml"] Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.422487 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.422636 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trqrd\" (UniqueName: \"kubernetes.io/projected/33b5c66b-1a29-44b1-9863-08dc69965dac-kube-api-access-trqrd\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.422721 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.525152 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.525308 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trqrd\" (UniqueName: \"kubernetes.io/projected/33b5c66b-1a29-44b1-9863-08dc69965dac-kube-api-access-trqrd\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.525399 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.529192 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.529357 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.551759 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trqrd\" (UniqueName: \"kubernetes.io/projected/33b5c66b-1a29-44b1-9863-08dc69965dac-kube-api-access-trqrd\") pod \"ssh-known-hosts-edpm-deployment-c7hml\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:56 crc kubenswrapper[4957]: I1206 06:17:56.712899 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:17:57 crc kubenswrapper[4957]: I1206 06:17:57.248865 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c7hml"] Dec 06 06:17:58 crc kubenswrapper[4957]: I1206 06:17:58.260000 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" event={"ID":"33b5c66b-1a29-44b1-9863-08dc69965dac","Type":"ContainerStarted","Data":"a6b72580073d5136327ebd4f6caecf5de2b6f2678afb90f7122b9d2eae13ebbd"} Dec 06 06:17:58 crc kubenswrapper[4957]: I1206 06:17:58.260555 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" event={"ID":"33b5c66b-1a29-44b1-9863-08dc69965dac","Type":"ContainerStarted","Data":"c7d7de198f8f1d3aced0d27e04a4aa4ecb446fe1c5cfec31b86b88d4475a1f33"} Dec 06 06:17:58 crc kubenswrapper[4957]: I1206 06:17:58.318208 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" podStartSLOduration=1.827227071 podStartE2EDuration="2.318182525s" podCreationTimestamp="2025-12-06 06:17:56 +0000 UTC" firstStartedPulling="2025-12-06 06:17:57.263500344 +0000 UTC m=+2316.913767976" lastFinishedPulling="2025-12-06 06:17:57.754455798 +0000 UTC m=+2317.404723430" observedRunningTime="2025-12-06 06:17:58.278601562 +0000 UTC m=+2317.928869214" watchObservedRunningTime="2025-12-06 06:17:58.318182525 +0000 UTC m=+2317.968450167" Dec 06 06:18:06 crc kubenswrapper[4957]: I1206 06:18:06.352721 4957 generic.go:334] "Generic (PLEG): container finished" podID="33b5c66b-1a29-44b1-9863-08dc69965dac" containerID="a6b72580073d5136327ebd4f6caecf5de2b6f2678afb90f7122b9d2eae13ebbd" exitCode=0 Dec 06 06:18:06 crc kubenswrapper[4957]: I1206 06:18:06.352808 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" event={"ID":"33b5c66b-1a29-44b1-9863-08dc69965dac","Type":"ContainerDied","Data":"a6b72580073d5136327ebd4f6caecf5de2b6f2678afb90f7122b9d2eae13ebbd"} Dec 06 06:18:06 crc kubenswrapper[4957]: I1206 06:18:06.663004 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:18:06 crc kubenswrapper[4957]: E1206 06:18:06.663349 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.776636 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.841927 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-ssh-key-openstack-edpm-ipam\") pod \"33b5c66b-1a29-44b1-9863-08dc69965dac\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.842024 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-inventory-0\") pod \"33b5c66b-1a29-44b1-9863-08dc69965dac\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.842154 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trqrd\" (UniqueName: \"kubernetes.io/projected/33b5c66b-1a29-44b1-9863-08dc69965dac-kube-api-access-trqrd\") pod \"33b5c66b-1a29-44b1-9863-08dc69965dac\" (UID: \"33b5c66b-1a29-44b1-9863-08dc69965dac\") " Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.848422 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b5c66b-1a29-44b1-9863-08dc69965dac-kube-api-access-trqrd" (OuterVolumeSpecName: "kube-api-access-trqrd") pod "33b5c66b-1a29-44b1-9863-08dc69965dac" (UID: "33b5c66b-1a29-44b1-9863-08dc69965dac"). InnerVolumeSpecName "kube-api-access-trqrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.869538 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "33b5c66b-1a29-44b1-9863-08dc69965dac" (UID: "33b5c66b-1a29-44b1-9863-08dc69965dac"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.871581 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "33b5c66b-1a29-44b1-9863-08dc69965dac" (UID: "33b5c66b-1a29-44b1-9863-08dc69965dac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.944319 4957 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.944352 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trqrd\" (UniqueName: \"kubernetes.io/projected/33b5c66b-1a29-44b1-9863-08dc69965dac-kube-api-access-trqrd\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:07 crc kubenswrapper[4957]: I1206 06:18:07.944363 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/33b5c66b-1a29-44b1-9863-08dc69965dac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.373378 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" event={"ID":"33b5c66b-1a29-44b1-9863-08dc69965dac","Type":"ContainerDied","Data":"c7d7de198f8f1d3aced0d27e04a4aa4ecb446fe1c5cfec31b86b88d4475a1f33"} Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.373432 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7d7de198f8f1d3aced0d27e04a4aa4ecb446fe1c5cfec31b86b88d4475a1f33" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.373516 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c7hml" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.453705 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd"] Dec 06 06:18:08 crc kubenswrapper[4957]: E1206 06:18:08.454408 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b5c66b-1a29-44b1-9863-08dc69965dac" containerName="ssh-known-hosts-edpm-deployment" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.454428 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b5c66b-1a29-44b1-9863-08dc69965dac" containerName="ssh-known-hosts-edpm-deployment" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.454682 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b5c66b-1a29-44b1-9863-08dc69965dac" containerName="ssh-known-hosts-edpm-deployment" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.455873 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.458476 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.458477 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.458648 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.458944 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.462496 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd"] Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.555290 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.555780 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps5kt\" (UniqueName: \"kubernetes.io/projected/9ca55297-0691-458e-85c2-a408a031c485-kube-api-access-ps5kt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.555873 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.657677 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps5kt\" (UniqueName: \"kubernetes.io/projected/9ca55297-0691-458e-85c2-a408a031c485-kube-api-access-ps5kt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.657732 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.657791 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.662094 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.662506 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.681472 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps5kt\" (UniqueName: \"kubernetes.io/projected/9ca55297-0691-458e-85c2-a408a031c485-kube-api-access-ps5kt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rk8hd\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:08 crc kubenswrapper[4957]: I1206 06:18:08.788587 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:09 crc kubenswrapper[4957]: I1206 06:18:09.322123 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd"] Dec 06 06:18:09 crc kubenswrapper[4957]: I1206 06:18:09.381641 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" event={"ID":"9ca55297-0691-458e-85c2-a408a031c485","Type":"ContainerStarted","Data":"c9eb38da09d374ea162ce4ecf5fc2ee2e81c345c59f30662563feae77601a24e"} Dec 06 06:18:10 crc kubenswrapper[4957]: I1206 06:18:10.392979 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" event={"ID":"9ca55297-0691-458e-85c2-a408a031c485","Type":"ContainerStarted","Data":"e6a52b5c768effa7a7096152d123ceb85b974aeedd0669d91e356f97f89d1c4c"} Dec 06 06:18:10 crc kubenswrapper[4957]: I1206 06:18:10.411992 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" podStartSLOduration=2.007412556 podStartE2EDuration="2.41197237s" podCreationTimestamp="2025-12-06 06:18:08 +0000 UTC" firstStartedPulling="2025-12-06 06:18:09.326730598 +0000 UTC m=+2328.976998230" lastFinishedPulling="2025-12-06 06:18:09.731290372 +0000 UTC m=+2329.381558044" observedRunningTime="2025-12-06 06:18:10.408902658 +0000 UTC m=+2330.059170320" watchObservedRunningTime="2025-12-06 06:18:10.41197237 +0000 UTC m=+2330.062240032" Dec 06 06:18:18 crc kubenswrapper[4957]: I1206 06:18:18.468952 4957 generic.go:334] "Generic (PLEG): container finished" podID="9ca55297-0691-458e-85c2-a408a031c485" containerID="e6a52b5c768effa7a7096152d123ceb85b974aeedd0669d91e356f97f89d1c4c" exitCode=0 Dec 06 06:18:18 crc kubenswrapper[4957]: I1206 06:18:18.469040 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" event={"ID":"9ca55297-0691-458e-85c2-a408a031c485","Type":"ContainerDied","Data":"e6a52b5c768effa7a7096152d123ceb85b974aeedd0669d91e356f97f89d1c4c"} Dec 06 06:18:19 crc kubenswrapper[4957]: I1206 06:18:19.907147 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:19 crc kubenswrapper[4957]: I1206 06:18:19.995677 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-ssh-key\") pod \"9ca55297-0691-458e-85c2-a408a031c485\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " Dec 06 06:18:19 crc kubenswrapper[4957]: I1206 06:18:19.995902 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-inventory\") pod \"9ca55297-0691-458e-85c2-a408a031c485\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " Dec 06 06:18:19 crc kubenswrapper[4957]: I1206 06:18:19.995954 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps5kt\" (UniqueName: \"kubernetes.io/projected/9ca55297-0691-458e-85c2-a408a031c485-kube-api-access-ps5kt\") pod \"9ca55297-0691-458e-85c2-a408a031c485\" (UID: \"9ca55297-0691-458e-85c2-a408a031c485\") " Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.001355 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ca55297-0691-458e-85c2-a408a031c485-kube-api-access-ps5kt" (OuterVolumeSpecName: "kube-api-access-ps5kt") pod "9ca55297-0691-458e-85c2-a408a031c485" (UID: "9ca55297-0691-458e-85c2-a408a031c485"). InnerVolumeSpecName "kube-api-access-ps5kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.024286 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9ca55297-0691-458e-85c2-a408a031c485" (UID: "9ca55297-0691-458e-85c2-a408a031c485"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.031661 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-inventory" (OuterVolumeSpecName: "inventory") pod "9ca55297-0691-458e-85c2-a408a031c485" (UID: "9ca55297-0691-458e-85c2-a408a031c485"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.098536 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.098578 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ca55297-0691-458e-85c2-a408a031c485-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.098589 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps5kt\" (UniqueName: \"kubernetes.io/projected/9ca55297-0691-458e-85c2-a408a031c485-kube-api-access-ps5kt\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.490310 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" event={"ID":"9ca55297-0691-458e-85c2-a408a031c485","Type":"ContainerDied","Data":"c9eb38da09d374ea162ce4ecf5fc2ee2e81c345c59f30662563feae77601a24e"} Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.490348 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9eb38da09d374ea162ce4ecf5fc2ee2e81c345c59f30662563feae77601a24e" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.490434 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rk8hd" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.599969 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv"] Dec 06 06:18:20 crc kubenswrapper[4957]: E1206 06:18:20.600489 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ca55297-0691-458e-85c2-a408a031c485" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.600518 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ca55297-0691-458e-85c2-a408a031c485" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.600781 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ca55297-0691-458e-85c2-a408a031c485" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.601600 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.604617 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.604761 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.604830 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.605070 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.613339 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv"] Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.710440 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvcdr\" (UniqueName: \"kubernetes.io/projected/101a78c1-076b-41be-a2bf-04f2f987924b-kube-api-access-mvcdr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.710490 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.710889 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.813205 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.813689 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvcdr\" (UniqueName: \"kubernetes.io/projected/101a78c1-076b-41be-a2bf-04f2f987924b-kube-api-access-mvcdr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.813726 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.816568 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.816815 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.829378 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.829595 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.837597 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvcdr\" (UniqueName: \"kubernetes.io/projected/101a78c1-076b-41be-a2bf-04f2f987924b-kube-api-access-mvcdr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.933241 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:18:20 crc kubenswrapper[4957]: I1206 06:18:20.942091 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:21 crc kubenswrapper[4957]: I1206 06:18:21.520272 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv"] Dec 06 06:18:21 crc kubenswrapper[4957]: I1206 06:18:21.662499 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:18:21 crc kubenswrapper[4957]: E1206 06:18:21.662780 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:18:22 crc kubenswrapper[4957]: I1206 06:18:22.346809 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:18:22 crc kubenswrapper[4957]: I1206 06:18:22.506924 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" event={"ID":"101a78c1-076b-41be-a2bf-04f2f987924b","Type":"ContainerStarted","Data":"5a5914e22380efd1cd4e80514402dbc46878032e5c1a39c9eee9eae6da4f7d7a"} Dec 06 06:18:23 crc kubenswrapper[4957]: I1206 06:18:23.514895 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" event={"ID":"101a78c1-076b-41be-a2bf-04f2f987924b","Type":"ContainerStarted","Data":"3cabd6b917f77abc0887cd46d775b246a7543ed3150a2dd5067511b5247e6918"} Dec 06 06:18:33 crc kubenswrapper[4957]: I1206 06:18:33.662423 4957 generic.go:334] "Generic (PLEG): container finished" podID="101a78c1-076b-41be-a2bf-04f2f987924b" containerID="3cabd6b917f77abc0887cd46d775b246a7543ed3150a2dd5067511b5247e6918" exitCode=0 Dec 06 06:18:33 crc kubenswrapper[4957]: I1206 06:18:33.662887 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" event={"ID":"101a78c1-076b-41be-a2bf-04f2f987924b","Type":"ContainerDied","Data":"3cabd6b917f77abc0887cd46d775b246a7543ed3150a2dd5067511b5247e6918"} Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.196555 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.329337 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-ssh-key\") pod \"101a78c1-076b-41be-a2bf-04f2f987924b\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.329415 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvcdr\" (UniqueName: \"kubernetes.io/projected/101a78c1-076b-41be-a2bf-04f2f987924b-kube-api-access-mvcdr\") pod \"101a78c1-076b-41be-a2bf-04f2f987924b\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.329467 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-inventory\") pod \"101a78c1-076b-41be-a2bf-04f2f987924b\" (UID: \"101a78c1-076b-41be-a2bf-04f2f987924b\") " Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.335159 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/101a78c1-076b-41be-a2bf-04f2f987924b-kube-api-access-mvcdr" (OuterVolumeSpecName: "kube-api-access-mvcdr") pod "101a78c1-076b-41be-a2bf-04f2f987924b" (UID: "101a78c1-076b-41be-a2bf-04f2f987924b"). InnerVolumeSpecName "kube-api-access-mvcdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.356659 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "101a78c1-076b-41be-a2bf-04f2f987924b" (UID: "101a78c1-076b-41be-a2bf-04f2f987924b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.365627 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-inventory" (OuterVolumeSpecName: "inventory") pod "101a78c1-076b-41be-a2bf-04f2f987924b" (UID: "101a78c1-076b-41be-a2bf-04f2f987924b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.431859 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.431918 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvcdr\" (UniqueName: \"kubernetes.io/projected/101a78c1-076b-41be-a2bf-04f2f987924b-kube-api-access-mvcdr\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.431974 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/101a78c1-076b-41be-a2bf-04f2f987924b-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.662794 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:18:35 crc kubenswrapper[4957]: E1206 06:18:35.663265 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.681600 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" event={"ID":"101a78c1-076b-41be-a2bf-04f2f987924b","Type":"ContainerDied","Data":"5a5914e22380efd1cd4e80514402dbc46878032e5c1a39c9eee9eae6da4f7d7a"} Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.681649 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a5914e22380efd1cd4e80514402dbc46878032e5c1a39c9eee9eae6da4f7d7a" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.681649 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.776562 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz"] Dec 06 06:18:35 crc kubenswrapper[4957]: E1206 06:18:35.776929 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101a78c1-076b-41be-a2bf-04f2f987924b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.776947 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="101a78c1-076b-41be-a2bf-04f2f987924b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.777133 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="101a78c1-076b-41be-a2bf-04f2f987924b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.777732 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.784871 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.785236 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.785291 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.785438 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.785673 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.785792 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.785883 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.785967 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.808064 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz"] Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.943905 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944114 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944242 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944328 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944420 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944530 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944643 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944764 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944890 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.944982 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.945070 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnjc8\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-kube-api-access-cnjc8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.945149 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.945217 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:35 crc kubenswrapper[4957]: I1206 06:18:35.945293 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.046786 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.046911 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047146 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047221 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047255 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047281 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047322 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047405 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047440 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047476 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047556 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnjc8\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-kube-api-access-cnjc8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047594 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047619 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.047658 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.053239 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.053650 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.053797 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.054069 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.054299 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.054439 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.055026 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.055257 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.055361 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.055533 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.056522 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.057028 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.057207 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.066268 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnjc8\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-kube-api-access-cnjc8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.104735 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:18:36 crc kubenswrapper[4957]: I1206 06:18:36.718265 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz"] Dec 06 06:18:36 crc kubenswrapper[4957]: W1206 06:18:36.735370 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26574c15_db39_470e_b739_cf03859b46aa.slice/crio-0a830302adc8d01236bb057f8d1d52c7fda0a1f68006fc28b9094adf70e2b3ba WatchSource:0}: Error finding container 0a830302adc8d01236bb057f8d1d52c7fda0a1f68006fc28b9094adf70e2b3ba: Status 404 returned error can't find the container with id 0a830302adc8d01236bb057f8d1d52c7fda0a1f68006fc28b9094adf70e2b3ba Dec 06 06:18:37 crc kubenswrapper[4957]: I1206 06:18:37.706244 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" event={"ID":"26574c15-db39-470e-b739-cf03859b46aa","Type":"ContainerStarted","Data":"1a0a3e80c40da3327b598b3cd4da0d4239f4088f957d4616a09564ff05813cb3"} Dec 06 06:18:37 crc kubenswrapper[4957]: I1206 06:18:37.706906 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" event={"ID":"26574c15-db39-470e-b739-cf03859b46aa","Type":"ContainerStarted","Data":"0a830302adc8d01236bb057f8d1d52c7fda0a1f68006fc28b9094adf70e2b3ba"} Dec 06 06:18:37 crc kubenswrapper[4957]: I1206 06:18:37.752548 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" podStartSLOduration=2.105664728 podStartE2EDuration="2.752523598s" podCreationTimestamp="2025-12-06 06:18:35 +0000 UTC" firstStartedPulling="2025-12-06 06:18:36.736867954 +0000 UTC m=+2356.387135586" lastFinishedPulling="2025-12-06 06:18:37.383726814 +0000 UTC m=+2357.033994456" observedRunningTime="2025-12-06 06:18:37.747797831 +0000 UTC m=+2357.398065483" watchObservedRunningTime="2025-12-06 06:18:37.752523598 +0000 UTC m=+2357.402791230" Dec 06 06:18:46 crc kubenswrapper[4957]: I1206 06:18:46.665656 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:18:46 crc kubenswrapper[4957]: E1206 06:18:46.684207 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:19:00 crc kubenswrapper[4957]: I1206 06:19:00.676346 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:19:00 crc kubenswrapper[4957]: E1206 06:19:00.677157 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:19:15 crc kubenswrapper[4957]: I1206 06:19:15.662917 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:19:15 crc kubenswrapper[4957]: E1206 06:19:15.664029 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:19:17 crc kubenswrapper[4957]: I1206 06:19:17.060094 4957 generic.go:334] "Generic (PLEG): container finished" podID="26574c15-db39-470e-b739-cf03859b46aa" containerID="1a0a3e80c40da3327b598b3cd4da0d4239f4088f957d4616a09564ff05813cb3" exitCode=0 Dec 06 06:19:17 crc kubenswrapper[4957]: I1206 06:19:17.060352 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" event={"ID":"26574c15-db39-470e-b739-cf03859b46aa","Type":"ContainerDied","Data":"1a0a3e80c40da3327b598b3cd4da0d4239f4088f957d4616a09564ff05813cb3"} Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.480544 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562152 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnjc8\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-kube-api-access-cnjc8\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562509 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-libvirt-combined-ca-bundle\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562645 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562678 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562697 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ovn-combined-ca-bundle\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562718 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ssh-key\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562744 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-telemetry-combined-ca-bundle\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562774 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-repo-setup-combined-ca-bundle\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562816 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-neutron-metadata-combined-ca-bundle\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562865 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562883 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-inventory\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562901 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.562970 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-nova-combined-ca-bundle\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.563005 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-bootstrap-combined-ca-bundle\") pod \"26574c15-db39-470e-b739-cf03859b46aa\" (UID: \"26574c15-db39-470e-b739-cf03859b46aa\") " Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.570094 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.574395 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.574478 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.574570 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.574643 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.574789 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.574698 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.574767 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.575239 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.575447 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-kube-api-access-cnjc8" (OuterVolumeSpecName: "kube-api-access-cnjc8") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "kube-api-access-cnjc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.575862 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.576659 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.605544 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-inventory" (OuterVolumeSpecName: "inventory") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.617377 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "26574c15-db39-470e-b739-cf03859b46aa" (UID: "26574c15-db39-470e-b739-cf03859b46aa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.664935 4957 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.664972 4957 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.664988 4957 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665004 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665015 4957 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665028 4957 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665041 4957 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665055 4957 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665070 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665082 4957 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665094 4957 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665106 4957 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665118 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnjc8\" (UniqueName: \"kubernetes.io/projected/26574c15-db39-470e-b739-cf03859b46aa-kube-api-access-cnjc8\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:18 crc kubenswrapper[4957]: I1206 06:19:18.665130 4957 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26574c15-db39-470e-b739-cf03859b46aa-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.083881 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" event={"ID":"26574c15-db39-470e-b739-cf03859b46aa","Type":"ContainerDied","Data":"0a830302adc8d01236bb057f8d1d52c7fda0a1f68006fc28b9094adf70e2b3ba"} Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.083926 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a830302adc8d01236bb057f8d1d52c7fda0a1f68006fc28b9094adf70e2b3ba" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.083943 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.236347 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh"] Dec 06 06:19:19 crc kubenswrapper[4957]: E1206 06:19:19.237748 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26574c15-db39-470e-b739-cf03859b46aa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.237857 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="26574c15-db39-470e-b739-cf03859b46aa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.241698 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="26574c15-db39-470e-b739-cf03859b46aa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.242608 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.256687 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.256841 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.256760 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.257960 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.258351 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.265695 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh"] Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.276768 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.276860 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m8dc\" (UniqueName: \"kubernetes.io/projected/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-kube-api-access-2m8dc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.276893 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.276936 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.277013 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.379050 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.379233 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.379262 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m8dc\" (UniqueName: \"kubernetes.io/projected/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-kube-api-access-2m8dc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.379281 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.379313 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.380573 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.383631 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.383726 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.383948 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.400149 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m8dc\" (UniqueName: \"kubernetes.io/projected/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-kube-api-access-2m8dc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8bwfh\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:19 crc kubenswrapper[4957]: I1206 06:19:19.573076 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:19:20 crc kubenswrapper[4957]: I1206 06:19:20.162906 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh"] Dec 06 06:19:20 crc kubenswrapper[4957]: W1206 06:19:20.164777 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eaa20e5_44dd_418a_94c2_fac0a16ef54a.slice/crio-c681fc2d3fdcbc73ff9026509bcb940d442a3f9ddd81dc64f38daee08b44bfde WatchSource:0}: Error finding container c681fc2d3fdcbc73ff9026509bcb940d442a3f9ddd81dc64f38daee08b44bfde: Status 404 returned error can't find the container with id c681fc2d3fdcbc73ff9026509bcb940d442a3f9ddd81dc64f38daee08b44bfde Dec 06 06:19:20 crc kubenswrapper[4957]: I1206 06:19:20.167395 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:19:21 crc kubenswrapper[4957]: I1206 06:19:21.105890 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" event={"ID":"1eaa20e5-44dd-418a-94c2-fac0a16ef54a","Type":"ContainerStarted","Data":"42390a2dbcce226d6fd4f101a1731ba7a4dc207a21f7ddfbb70a4643e13c7f82"} Dec 06 06:19:21 crc kubenswrapper[4957]: I1206 06:19:21.106228 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" event={"ID":"1eaa20e5-44dd-418a-94c2-fac0a16ef54a","Type":"ContainerStarted","Data":"c681fc2d3fdcbc73ff9026509bcb940d442a3f9ddd81dc64f38daee08b44bfde"} Dec 06 06:19:21 crc kubenswrapper[4957]: I1206 06:19:21.136674 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" podStartSLOduration=1.713105764 podStartE2EDuration="2.136650657s" podCreationTimestamp="2025-12-06 06:19:19 +0000 UTC" firstStartedPulling="2025-12-06 06:19:20.167195944 +0000 UTC m=+2399.817463576" lastFinishedPulling="2025-12-06 06:19:20.590740837 +0000 UTC m=+2400.241008469" observedRunningTime="2025-12-06 06:19:21.131165509 +0000 UTC m=+2400.781433141" watchObservedRunningTime="2025-12-06 06:19:21.136650657 +0000 UTC m=+2400.786918299" Dec 06 06:19:27 crc kubenswrapper[4957]: I1206 06:19:27.662083 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:19:27 crc kubenswrapper[4957]: E1206 06:19:27.662935 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:19:38 crc kubenswrapper[4957]: I1206 06:19:38.662646 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:19:38 crc kubenswrapper[4957]: E1206 06:19:38.663527 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:19:50 crc kubenswrapper[4957]: I1206 06:19:50.669543 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:19:50 crc kubenswrapper[4957]: E1206 06:19:50.684873 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:20:04 crc kubenswrapper[4957]: I1206 06:20:04.662648 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:20:04 crc kubenswrapper[4957]: E1206 06:20:04.663477 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:20:17 crc kubenswrapper[4957]: I1206 06:20:17.663459 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:20:17 crc kubenswrapper[4957]: E1206 06:20:17.664628 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:20:24 crc kubenswrapper[4957]: I1206 06:20:24.727636 4957 generic.go:334] "Generic (PLEG): container finished" podID="1eaa20e5-44dd-418a-94c2-fac0a16ef54a" containerID="42390a2dbcce226d6fd4f101a1731ba7a4dc207a21f7ddfbb70a4643e13c7f82" exitCode=0 Dec 06 06:20:24 crc kubenswrapper[4957]: I1206 06:20:24.727726 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" event={"ID":"1eaa20e5-44dd-418a-94c2-fac0a16ef54a","Type":"ContainerDied","Data":"42390a2dbcce226d6fd4f101a1731ba7a4dc207a21f7ddfbb70a4643e13c7f82"} Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.192407 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.322363 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ssh-key\") pod \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.322531 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-inventory\") pod \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.322563 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m8dc\" (UniqueName: \"kubernetes.io/projected/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-kube-api-access-2m8dc\") pod \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.322583 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovncontroller-config-0\") pod \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.322631 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovn-combined-ca-bundle\") pod \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\" (UID: \"1eaa20e5-44dd-418a-94c2-fac0a16ef54a\") " Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.328563 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1eaa20e5-44dd-418a-94c2-fac0a16ef54a" (UID: "1eaa20e5-44dd-418a-94c2-fac0a16ef54a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.333187 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-kube-api-access-2m8dc" (OuterVolumeSpecName: "kube-api-access-2m8dc") pod "1eaa20e5-44dd-418a-94c2-fac0a16ef54a" (UID: "1eaa20e5-44dd-418a-94c2-fac0a16ef54a"). InnerVolumeSpecName "kube-api-access-2m8dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.353715 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-inventory" (OuterVolumeSpecName: "inventory") pod "1eaa20e5-44dd-418a-94c2-fac0a16ef54a" (UID: "1eaa20e5-44dd-418a-94c2-fac0a16ef54a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.375646 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1eaa20e5-44dd-418a-94c2-fac0a16ef54a" (UID: "1eaa20e5-44dd-418a-94c2-fac0a16ef54a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.379666 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "1eaa20e5-44dd-418a-94c2-fac0a16ef54a" (UID: "1eaa20e5-44dd-418a-94c2-fac0a16ef54a"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.424951 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.424979 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m8dc\" (UniqueName: \"kubernetes.io/projected/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-kube-api-access-2m8dc\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.424991 4957 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.424999 4957 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.425007 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eaa20e5-44dd-418a-94c2-fac0a16ef54a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.751097 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" event={"ID":"1eaa20e5-44dd-418a-94c2-fac0a16ef54a","Type":"ContainerDied","Data":"c681fc2d3fdcbc73ff9026509bcb940d442a3f9ddd81dc64f38daee08b44bfde"} Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.751180 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c681fc2d3fdcbc73ff9026509bcb940d442a3f9ddd81dc64f38daee08b44bfde" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.751248 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8bwfh" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.835781 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n"] Dec 06 06:20:26 crc kubenswrapper[4957]: E1206 06:20:26.836273 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eaa20e5-44dd-418a-94c2-fac0a16ef54a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.836308 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eaa20e5-44dd-418a-94c2-fac0a16ef54a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.836523 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eaa20e5-44dd-418a-94c2-fac0a16ef54a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.837252 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.838983 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.839330 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.839440 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.839499 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.840215 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.841705 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.848807 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n"] Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.935376 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.935425 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.935460 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.935650 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.935735 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:26 crc kubenswrapper[4957]: I1206 06:20:26.935936 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grtbb\" (UniqueName: \"kubernetes.io/projected/a0ec714a-36ad-46f4-aeca-a912266812f2-kube-api-access-grtbb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.037687 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.037938 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.038039 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.038182 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.038316 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.038899 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grtbb\" (UniqueName: \"kubernetes.io/projected/a0ec714a-36ad-46f4-aeca-a912266812f2-kube-api-access-grtbb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.043165 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.043356 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.044261 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.044413 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.053776 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.055215 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grtbb\" (UniqueName: \"kubernetes.io/projected/a0ec714a-36ad-46f4-aeca-a912266812f2-kube-api-access-grtbb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.158278 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.719271 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n"] Dec 06 06:20:27 crc kubenswrapper[4957]: I1206 06:20:27.772024 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" event={"ID":"a0ec714a-36ad-46f4-aeca-a912266812f2","Type":"ContainerStarted","Data":"f22cfd8d53304eedd7e4d6fa09a4978e9c056129b3d91304a63190e62640db96"} Dec 06 06:20:28 crc kubenswrapper[4957]: I1206 06:20:28.785118 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" event={"ID":"a0ec714a-36ad-46f4-aeca-a912266812f2","Type":"ContainerStarted","Data":"59c37e9bec8b76185591222faa11b833742c54701cfebaf9abfd9c64b3223773"} Dec 06 06:20:28 crc kubenswrapper[4957]: I1206 06:20:28.820183 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" podStartSLOduration=2.359265006 podStartE2EDuration="2.820159826s" podCreationTimestamp="2025-12-06 06:20:26 +0000 UTC" firstStartedPulling="2025-12-06 06:20:27.731923661 +0000 UTC m=+2467.382191283" lastFinishedPulling="2025-12-06 06:20:28.192818431 +0000 UTC m=+2467.843086103" observedRunningTime="2025-12-06 06:20:28.804011637 +0000 UTC m=+2468.454279309" watchObservedRunningTime="2025-12-06 06:20:28.820159826 +0000 UTC m=+2468.470427478" Dec 06 06:20:32 crc kubenswrapper[4957]: I1206 06:20:32.662934 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:20:32 crc kubenswrapper[4957]: E1206 06:20:32.663467 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.208710 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fmfws"] Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.211346 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.233441 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fmfws"] Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.261610 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-catalog-content\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.262082 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-utilities\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.262174 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g5xm\" (UniqueName: \"kubernetes.io/projected/e654b71f-f8db-42a6-8a3b-42e4e88651fb-kube-api-access-8g5xm\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.364683 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-catalog-content\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.365393 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-catalog-content\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.365611 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-utilities\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.365733 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g5xm\" (UniqueName: \"kubernetes.io/projected/e654b71f-f8db-42a6-8a3b-42e4e88651fb-kube-api-access-8g5xm\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.365995 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-utilities\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.389219 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g5xm\" (UniqueName: \"kubernetes.io/projected/e654b71f-f8db-42a6-8a3b-42e4e88651fb-kube-api-access-8g5xm\") pod \"certified-operators-fmfws\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:42 crc kubenswrapper[4957]: I1206 06:20:42.544658 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:43 crc kubenswrapper[4957]: I1206 06:20:43.047570 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fmfws"] Dec 06 06:20:43 crc kubenswrapper[4957]: W1206 06:20:43.047978 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode654b71f_f8db_42a6_8a3b_42e4e88651fb.slice/crio-900bcd6e0c5451910b47311c85f4885357764c59058e57c961898ba2cf5bf429 WatchSource:0}: Error finding container 900bcd6e0c5451910b47311c85f4885357764c59058e57c961898ba2cf5bf429: Status 404 returned error can't find the container with id 900bcd6e0c5451910b47311c85f4885357764c59058e57c961898ba2cf5bf429 Dec 06 06:20:43 crc kubenswrapper[4957]: I1206 06:20:43.934207 4957 generic.go:334] "Generic (PLEG): container finished" podID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerID="61d300446c5870f8c5226ec0c7a1334a51a186735fa6b5f20596f537d06e0dcb" exitCode=0 Dec 06 06:20:43 crc kubenswrapper[4957]: I1206 06:20:43.934295 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmfws" event={"ID":"e654b71f-f8db-42a6-8a3b-42e4e88651fb","Type":"ContainerDied","Data":"61d300446c5870f8c5226ec0c7a1334a51a186735fa6b5f20596f537d06e0dcb"} Dec 06 06:20:43 crc kubenswrapper[4957]: I1206 06:20:43.934616 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmfws" event={"ID":"e654b71f-f8db-42a6-8a3b-42e4e88651fb","Type":"ContainerStarted","Data":"900bcd6e0c5451910b47311c85f4885357764c59058e57c961898ba2cf5bf429"} Dec 06 06:20:44 crc kubenswrapper[4957]: I1206 06:20:44.945408 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmfws" event={"ID":"e654b71f-f8db-42a6-8a3b-42e4e88651fb","Type":"ContainerStarted","Data":"e5d4b0b651f286018f708af2ba1595fe9112bbe0a5020579fe166917d5ae57c4"} Dec 06 06:20:45 crc kubenswrapper[4957]: I1206 06:20:45.956051 4957 generic.go:334] "Generic (PLEG): container finished" podID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerID="e5d4b0b651f286018f708af2ba1595fe9112bbe0a5020579fe166917d5ae57c4" exitCode=0 Dec 06 06:20:45 crc kubenswrapper[4957]: I1206 06:20:45.956131 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmfws" event={"ID":"e654b71f-f8db-42a6-8a3b-42e4e88651fb","Type":"ContainerDied","Data":"e5d4b0b651f286018f708af2ba1595fe9112bbe0a5020579fe166917d5ae57c4"} Dec 06 06:20:46 crc kubenswrapper[4957]: I1206 06:20:46.967423 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmfws" event={"ID":"e654b71f-f8db-42a6-8a3b-42e4e88651fb","Type":"ContainerStarted","Data":"7910f708460924456e6f50ae78bf4b466fbb22594c7397a280ecd1a46dbe9aa9"} Dec 06 06:20:46 crc kubenswrapper[4957]: I1206 06:20:46.991924 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fmfws" podStartSLOduration=2.578774572 podStartE2EDuration="4.991907066s" podCreationTimestamp="2025-12-06 06:20:42 +0000 UTC" firstStartedPulling="2025-12-06 06:20:43.936695157 +0000 UTC m=+2483.586962809" lastFinishedPulling="2025-12-06 06:20:46.349827661 +0000 UTC m=+2486.000095303" observedRunningTime="2025-12-06 06:20:46.990118077 +0000 UTC m=+2486.640385739" watchObservedRunningTime="2025-12-06 06:20:46.991907066 +0000 UTC m=+2486.642174688" Dec 06 06:20:47 crc kubenswrapper[4957]: I1206 06:20:47.662248 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:20:47 crc kubenswrapper[4957]: E1206 06:20:47.662592 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:20:52 crc kubenswrapper[4957]: I1206 06:20:52.545853 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:52 crc kubenswrapper[4957]: I1206 06:20:52.546494 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:52 crc kubenswrapper[4957]: I1206 06:20:52.597527 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:53 crc kubenswrapper[4957]: I1206 06:20:53.064396 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:53 crc kubenswrapper[4957]: I1206 06:20:53.121186 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fmfws"] Dec 06 06:20:55 crc kubenswrapper[4957]: I1206 06:20:55.032192 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fmfws" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="registry-server" containerID="cri-o://7910f708460924456e6f50ae78bf4b466fbb22594c7397a280ecd1a46dbe9aa9" gracePeriod=2 Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.044994 4957 generic.go:334] "Generic (PLEG): container finished" podID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerID="7910f708460924456e6f50ae78bf4b466fbb22594c7397a280ecd1a46dbe9aa9" exitCode=0 Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.045035 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmfws" event={"ID":"e654b71f-f8db-42a6-8a3b-42e4e88651fb","Type":"ContainerDied","Data":"7910f708460924456e6f50ae78bf4b466fbb22594c7397a280ecd1a46dbe9aa9"} Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.598039 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.755127 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g5xm\" (UniqueName: \"kubernetes.io/projected/e654b71f-f8db-42a6-8a3b-42e4e88651fb-kube-api-access-8g5xm\") pod \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.755714 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-utilities\") pod \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.755963 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-catalog-content\") pod \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\" (UID: \"e654b71f-f8db-42a6-8a3b-42e4e88651fb\") " Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.756429 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-utilities" (OuterVolumeSpecName: "utilities") pod "e654b71f-f8db-42a6-8a3b-42e4e88651fb" (UID: "e654b71f-f8db-42a6-8a3b-42e4e88651fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.760702 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e654b71f-f8db-42a6-8a3b-42e4e88651fb-kube-api-access-8g5xm" (OuterVolumeSpecName: "kube-api-access-8g5xm") pod "e654b71f-f8db-42a6-8a3b-42e4e88651fb" (UID: "e654b71f-f8db-42a6-8a3b-42e4e88651fb"). InnerVolumeSpecName "kube-api-access-8g5xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.804262 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e654b71f-f8db-42a6-8a3b-42e4e88651fb" (UID: "e654b71f-f8db-42a6-8a3b-42e4e88651fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.858113 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.858146 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e654b71f-f8db-42a6-8a3b-42e4e88651fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:56 crc kubenswrapper[4957]: I1206 06:20:56.858157 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g5xm\" (UniqueName: \"kubernetes.io/projected/e654b71f-f8db-42a6-8a3b-42e4e88651fb-kube-api-access-8g5xm\") on node \"crc\" DevicePath \"\"" Dec 06 06:20:57 crc kubenswrapper[4957]: I1206 06:20:57.057805 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmfws" event={"ID":"e654b71f-f8db-42a6-8a3b-42e4e88651fb","Type":"ContainerDied","Data":"900bcd6e0c5451910b47311c85f4885357764c59058e57c961898ba2cf5bf429"} Dec 06 06:20:57 crc kubenswrapper[4957]: I1206 06:20:57.057894 4957 scope.go:117] "RemoveContainer" containerID="7910f708460924456e6f50ae78bf4b466fbb22594c7397a280ecd1a46dbe9aa9" Dec 06 06:20:57 crc kubenswrapper[4957]: I1206 06:20:57.057912 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmfws" Dec 06 06:20:57 crc kubenswrapper[4957]: I1206 06:20:57.089673 4957 scope.go:117] "RemoveContainer" containerID="e5d4b0b651f286018f708af2ba1595fe9112bbe0a5020579fe166917d5ae57c4" Dec 06 06:20:57 crc kubenswrapper[4957]: I1206 06:20:57.098607 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fmfws"] Dec 06 06:20:57 crc kubenswrapper[4957]: I1206 06:20:57.106320 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fmfws"] Dec 06 06:20:57 crc kubenswrapper[4957]: I1206 06:20:57.113973 4957 scope.go:117] "RemoveContainer" containerID="61d300446c5870f8c5226ec0c7a1334a51a186735fa6b5f20596f537d06e0dcb" Dec 06 06:20:58 crc kubenswrapper[4957]: I1206 06:20:58.673883 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" path="/var/lib/kubelet/pods/e654b71f-f8db-42a6-8a3b-42e4e88651fb/volumes" Dec 06 06:21:02 crc kubenswrapper[4957]: I1206 06:21:02.662292 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:21:02 crc kubenswrapper[4957]: E1206 06:21:02.663071 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:21:13 crc kubenswrapper[4957]: I1206 06:21:13.664148 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:21:13 crc kubenswrapper[4957]: E1206 06:21:13.664944 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:21:17 crc kubenswrapper[4957]: I1206 06:21:17.265592 4957 generic.go:334] "Generic (PLEG): container finished" podID="a0ec714a-36ad-46f4-aeca-a912266812f2" containerID="59c37e9bec8b76185591222faa11b833742c54701cfebaf9abfd9c64b3223773" exitCode=0 Dec 06 06:21:17 crc kubenswrapper[4957]: I1206 06:21:17.265669 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" event={"ID":"a0ec714a-36ad-46f4-aeca-a912266812f2","Type":"ContainerDied","Data":"59c37e9bec8b76185591222faa11b833742c54701cfebaf9abfd9c64b3223773"} Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.685725 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.789510 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grtbb\" (UniqueName: \"kubernetes.io/projected/a0ec714a-36ad-46f4-aeca-a912266812f2-kube-api-access-grtbb\") pod \"a0ec714a-36ad-46f4-aeca-a912266812f2\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.789685 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-nova-metadata-neutron-config-0\") pod \"a0ec714a-36ad-46f4-aeca-a912266812f2\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.789739 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-ssh-key\") pod \"a0ec714a-36ad-46f4-aeca-a912266812f2\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.789789 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-metadata-combined-ca-bundle\") pod \"a0ec714a-36ad-46f4-aeca-a912266812f2\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.789804 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a0ec714a-36ad-46f4-aeca-a912266812f2\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.789844 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-inventory\") pod \"a0ec714a-36ad-46f4-aeca-a912266812f2\" (UID: \"a0ec714a-36ad-46f4-aeca-a912266812f2\") " Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.801940 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a0ec714a-36ad-46f4-aeca-a912266812f2" (UID: "a0ec714a-36ad-46f4-aeca-a912266812f2"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.804134 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ec714a-36ad-46f4-aeca-a912266812f2-kube-api-access-grtbb" (OuterVolumeSpecName: "kube-api-access-grtbb") pod "a0ec714a-36ad-46f4-aeca-a912266812f2" (UID: "a0ec714a-36ad-46f4-aeca-a912266812f2"). InnerVolumeSpecName "kube-api-access-grtbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.817219 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-inventory" (OuterVolumeSpecName: "inventory") pod "a0ec714a-36ad-46f4-aeca-a912266812f2" (UID: "a0ec714a-36ad-46f4-aeca-a912266812f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.817942 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a0ec714a-36ad-46f4-aeca-a912266812f2" (UID: "a0ec714a-36ad-46f4-aeca-a912266812f2"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.833930 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a0ec714a-36ad-46f4-aeca-a912266812f2" (UID: "a0ec714a-36ad-46f4-aeca-a912266812f2"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.837366 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a0ec714a-36ad-46f4-aeca-a912266812f2" (UID: "a0ec714a-36ad-46f4-aeca-a912266812f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.891587 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grtbb\" (UniqueName: \"kubernetes.io/projected/a0ec714a-36ad-46f4-aeca-a912266812f2-kube-api-access-grtbb\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.891887 4957 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.891959 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.892076 4957 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.892149 4957 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:18 crc kubenswrapper[4957]: I1206 06:21:18.892205 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0ec714a-36ad-46f4-aeca-a912266812f2-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.283102 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" event={"ID":"a0ec714a-36ad-46f4-aeca-a912266812f2","Type":"ContainerDied","Data":"f22cfd8d53304eedd7e4d6fa09a4978e9c056129b3d91304a63190e62640db96"} Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.283144 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f22cfd8d53304eedd7e4d6fa09a4978e9c056129b3d91304a63190e62640db96" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.283170 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.421012 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz"] Dec 06 06:21:19 crc kubenswrapper[4957]: E1206 06:21:19.421382 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="extract-utilities" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.421403 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="extract-utilities" Dec 06 06:21:19 crc kubenswrapper[4957]: E1206 06:21:19.421440 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ec714a-36ad-46f4-aeca-a912266812f2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.421449 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ec714a-36ad-46f4-aeca-a912266812f2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 06:21:19 crc kubenswrapper[4957]: E1206 06:21:19.421459 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="registry-server" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.421464 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="registry-server" Dec 06 06:21:19 crc kubenswrapper[4957]: E1206 06:21:19.421474 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="extract-content" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.421480 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="extract-content" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.421639 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e654b71f-f8db-42a6-8a3b-42e4e88651fb" containerName="registry-server" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.421663 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ec714a-36ad-46f4-aeca-a912266812f2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.422310 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.429552 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.429577 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.429954 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.429996 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.430259 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.441665 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz"] Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.504087 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6fls\" (UniqueName: \"kubernetes.io/projected/5df54304-6d88-4027-ae21-40aaa328470f-kube-api-access-x6fls\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.504132 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.504308 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.504407 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.504672 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.606550 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6fls\" (UniqueName: \"kubernetes.io/projected/5df54304-6d88-4027-ae21-40aaa328470f-kube-api-access-x6fls\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.606620 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.606693 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.606755 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.606925 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.612131 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.612416 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.621025 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.621926 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.641260 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6fls\" (UniqueName: \"kubernetes.io/projected/5df54304-6d88-4027-ae21-40aaa328470f-kube-api-access-x6fls\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:19 crc kubenswrapper[4957]: I1206 06:21:19.738896 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:21:20 crc kubenswrapper[4957]: I1206 06:21:20.283973 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz"] Dec 06 06:21:20 crc kubenswrapper[4957]: W1206 06:21:20.286268 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5df54304_6d88_4027_ae21_40aaa328470f.slice/crio-33838172f016da658ed6ec623e128ab0d2ad389e4376c83d9bb769cc5d9debea WatchSource:0}: Error finding container 33838172f016da658ed6ec623e128ab0d2ad389e4376c83d9bb769cc5d9debea: Status 404 returned error can't find the container with id 33838172f016da658ed6ec623e128ab0d2ad389e4376c83d9bb769cc5d9debea Dec 06 06:21:20 crc kubenswrapper[4957]: I1206 06:21:20.758067 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:21:21 crc kubenswrapper[4957]: I1206 06:21:21.306736 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" event={"ID":"5df54304-6d88-4027-ae21-40aaa328470f","Type":"ContainerStarted","Data":"e5acbcbe2dddfe1eb1024d16d744f1154a07a2717aa063088ce2b2ca63a1929a"} Dec 06 06:21:21 crc kubenswrapper[4957]: I1206 06:21:21.306786 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" event={"ID":"5df54304-6d88-4027-ae21-40aaa328470f","Type":"ContainerStarted","Data":"33838172f016da658ed6ec623e128ab0d2ad389e4376c83d9bb769cc5d9debea"} Dec 06 06:21:26 crc kubenswrapper[4957]: I1206 06:21:26.662209 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:21:26 crc kubenswrapper[4957]: E1206 06:21:26.662994 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:21:37 crc kubenswrapper[4957]: I1206 06:21:37.663112 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:21:37 crc kubenswrapper[4957]: E1206 06:21:37.664356 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:21:52 crc kubenswrapper[4957]: I1206 06:21:52.663077 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:21:52 crc kubenswrapper[4957]: E1206 06:21:52.663948 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:22:03 crc kubenswrapper[4957]: I1206 06:22:03.662220 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:22:03 crc kubenswrapper[4957]: E1206 06:22:03.663095 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:22:16 crc kubenswrapper[4957]: I1206 06:22:16.662793 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:22:17 crc kubenswrapper[4957]: I1206 06:22:17.865745 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"f17feb3bb43f81994797083be9593f7ff881e7b682b073303dbfac3dc781fc14"} Dec 06 06:22:17 crc kubenswrapper[4957]: I1206 06:22:17.885095 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" podStartSLOduration=58.417979447 podStartE2EDuration="58.885075206s" podCreationTimestamp="2025-12-06 06:21:19 +0000 UTC" firstStartedPulling="2025-12-06 06:21:20.28869753 +0000 UTC m=+2519.938965162" lastFinishedPulling="2025-12-06 06:21:20.755793289 +0000 UTC m=+2520.406060921" observedRunningTime="2025-12-06 06:21:21.326421121 +0000 UTC m=+2520.976688783" watchObservedRunningTime="2025-12-06 06:22:17.885075206 +0000 UTC m=+2577.535342838" Dec 06 06:24:40 crc kubenswrapper[4957]: I1206 06:24:40.180951 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:24:40 crc kubenswrapper[4957]: I1206 06:24:40.181575 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.088194 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-khx5x"] Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.092025 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.102973 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-khx5x"] Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.221261 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-utilities\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.221333 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-catalog-content\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.221983 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/559c6b9f-c9c4-4248-aff8-79a7daff9442-kube-api-access-r59mp\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.324023 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/559c6b9f-c9c4-4248-aff8-79a7daff9442-kube-api-access-r59mp\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.324093 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-utilities\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.324169 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-catalog-content\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.324694 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-utilities\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.324738 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-catalog-content\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.346348 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/559c6b9f-c9c4-4248-aff8-79a7daff9442-kube-api-access-r59mp\") pod \"community-operators-khx5x\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.420002 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:00 crc kubenswrapper[4957]: I1206 06:25:00.961227 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-khx5x"] Dec 06 06:25:01 crc kubenswrapper[4957]: I1206 06:25:01.408490 4957 generic.go:334] "Generic (PLEG): container finished" podID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerID="d13c1fa04295d8615bcf79373a2299a2a4ed78dfb7efa5bf2b5cb9fdb1bc4a06" exitCode=0 Dec 06 06:25:01 crc kubenswrapper[4957]: I1206 06:25:01.408550 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khx5x" event={"ID":"559c6b9f-c9c4-4248-aff8-79a7daff9442","Type":"ContainerDied","Data":"d13c1fa04295d8615bcf79373a2299a2a4ed78dfb7efa5bf2b5cb9fdb1bc4a06"} Dec 06 06:25:01 crc kubenswrapper[4957]: I1206 06:25:01.408791 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khx5x" event={"ID":"559c6b9f-c9c4-4248-aff8-79a7daff9442","Type":"ContainerStarted","Data":"53496ff3b42d06c41f9cf47ee4f45cc1e595679a2a1807aeb3bc10c51d01b2e8"} Dec 06 06:25:01 crc kubenswrapper[4957]: I1206 06:25:01.411013 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:25:02 crc kubenswrapper[4957]: I1206 06:25:02.420375 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khx5x" event={"ID":"559c6b9f-c9c4-4248-aff8-79a7daff9442","Type":"ContainerStarted","Data":"8b52aa6111a4455d2e8f4c7cc6fbb590268f06e02a38acf9f28f1e59b866b2cb"} Dec 06 06:25:03 crc kubenswrapper[4957]: I1206 06:25:03.430224 4957 generic.go:334] "Generic (PLEG): container finished" podID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerID="8b52aa6111a4455d2e8f4c7cc6fbb590268f06e02a38acf9f28f1e59b866b2cb" exitCode=0 Dec 06 06:25:03 crc kubenswrapper[4957]: I1206 06:25:03.430274 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khx5x" event={"ID":"559c6b9f-c9c4-4248-aff8-79a7daff9442","Type":"ContainerDied","Data":"8b52aa6111a4455d2e8f4c7cc6fbb590268f06e02a38acf9f28f1e59b866b2cb"} Dec 06 06:25:06 crc kubenswrapper[4957]: I1206 06:25:06.455676 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khx5x" event={"ID":"559c6b9f-c9c4-4248-aff8-79a7daff9442","Type":"ContainerStarted","Data":"3f7548394b9204d8bcaf9293f131877835153c08172a506c8f3917b4f51bc790"} Dec 06 06:25:06 crc kubenswrapper[4957]: I1206 06:25:06.480110 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-khx5x" podStartSLOduration=4.073556722 podStartE2EDuration="6.480077198s" podCreationTimestamp="2025-12-06 06:25:00 +0000 UTC" firstStartedPulling="2025-12-06 06:25:01.410706781 +0000 UTC m=+2741.060974413" lastFinishedPulling="2025-12-06 06:25:03.817227247 +0000 UTC m=+2743.467494889" observedRunningTime="2025-12-06 06:25:06.474548628 +0000 UTC m=+2746.124816290" watchObservedRunningTime="2025-12-06 06:25:06.480077198 +0000 UTC m=+2746.130344830" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.483432 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gp7zb"] Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.486157 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.499199 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gp7zb"] Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.676924 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvcwd\" (UniqueName: \"kubernetes.io/projected/a3f95673-ae9e-4ee6-9950-06ca90421e76-kube-api-access-kvcwd\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.677003 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-catalog-content\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.677070 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-utilities\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.778268 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-utilities\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.778441 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvcwd\" (UniqueName: \"kubernetes.io/projected/a3f95673-ae9e-4ee6-9950-06ca90421e76-kube-api-access-kvcwd\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.778467 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-catalog-content\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.778894 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-utilities\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.779316 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-catalog-content\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.797118 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvcwd\" (UniqueName: \"kubernetes.io/projected/a3f95673-ae9e-4ee6-9950-06ca90421e76-kube-api-access-kvcwd\") pod \"redhat-operators-gp7zb\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:07 crc kubenswrapper[4957]: I1206 06:25:07.804991 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:08 crc kubenswrapper[4957]: I1206 06:25:08.304071 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gp7zb"] Dec 06 06:25:08 crc kubenswrapper[4957]: W1206 06:25:08.314407 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3f95673_ae9e_4ee6_9950_06ca90421e76.slice/crio-77e63140b6591559fe3342573ff9be39cc4b3a990627d41f25c32f2a21f6f4bd WatchSource:0}: Error finding container 77e63140b6591559fe3342573ff9be39cc4b3a990627d41f25c32f2a21f6f4bd: Status 404 returned error can't find the container with id 77e63140b6591559fe3342573ff9be39cc4b3a990627d41f25c32f2a21f6f4bd Dec 06 06:25:08 crc kubenswrapper[4957]: I1206 06:25:08.472111 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gp7zb" event={"ID":"a3f95673-ae9e-4ee6-9950-06ca90421e76","Type":"ContainerStarted","Data":"77e63140b6591559fe3342573ff9be39cc4b3a990627d41f25c32f2a21f6f4bd"} Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.181276 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.181611 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.421132 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.421557 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.479846 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.534486 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.708370 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g6kvh"] Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.724600 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.724944 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6kvh"] Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.837468 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-utilities\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.837692 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75g5j\" (UniqueName: \"kubernetes.io/projected/09ce914e-ef64-46dd-8aad-eafc79d08d2f-kube-api-access-75g5j\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.838376 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-catalog-content\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.940732 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75g5j\" (UniqueName: \"kubernetes.io/projected/09ce914e-ef64-46dd-8aad-eafc79d08d2f-kube-api-access-75g5j\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.940849 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-catalog-content\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.940975 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-utilities\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.941444 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-utilities\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.942037 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-catalog-content\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:10 crc kubenswrapper[4957]: I1206 06:25:10.962283 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75g5j\" (UniqueName: \"kubernetes.io/projected/09ce914e-ef64-46dd-8aad-eafc79d08d2f-kube-api-access-75g5j\") pod \"redhat-marketplace-g6kvh\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:11 crc kubenswrapper[4957]: I1206 06:25:11.049965 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:11 crc kubenswrapper[4957]: I1206 06:25:11.299801 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6kvh"] Dec 06 06:25:11 crc kubenswrapper[4957]: I1206 06:25:11.495381 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6kvh" event={"ID":"09ce914e-ef64-46dd-8aad-eafc79d08d2f","Type":"ContainerStarted","Data":"333fe18f00772619bd0bbf21617b9908c552822a2129f1bc715bd2f64f21ffb2"} Dec 06 06:25:12 crc kubenswrapper[4957]: I1206 06:25:12.876079 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-khx5x"] Dec 06 06:25:12 crc kubenswrapper[4957]: I1206 06:25:12.876977 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-khx5x" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="registry-server" containerID="cri-o://3f7548394b9204d8bcaf9293f131877835153c08172a506c8f3917b4f51bc790" gracePeriod=2 Dec 06 06:25:14 crc kubenswrapper[4957]: I1206 06:25:14.531292 4957 generic.go:334] "Generic (PLEG): container finished" podID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerID="b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084" exitCode=0 Dec 06 06:25:14 crc kubenswrapper[4957]: I1206 06:25:14.531406 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6kvh" event={"ID":"09ce914e-ef64-46dd-8aad-eafc79d08d2f","Type":"ContainerDied","Data":"b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084"} Dec 06 06:25:14 crc kubenswrapper[4957]: I1206 06:25:14.537769 4957 generic.go:334] "Generic (PLEG): container finished" podID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerID="3f7548394b9204d8bcaf9293f131877835153c08172a506c8f3917b4f51bc790" exitCode=0 Dec 06 06:25:14 crc kubenswrapper[4957]: I1206 06:25:14.537847 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khx5x" event={"ID":"559c6b9f-c9c4-4248-aff8-79a7daff9442","Type":"ContainerDied","Data":"3f7548394b9204d8bcaf9293f131877835153c08172a506c8f3917b4f51bc790"} Dec 06 06:25:14 crc kubenswrapper[4957]: I1206 06:25:14.540298 4957 generic.go:334] "Generic (PLEG): container finished" podID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerID="fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613" exitCode=0 Dec 06 06:25:14 crc kubenswrapper[4957]: I1206 06:25:14.540339 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gp7zb" event={"ID":"a3f95673-ae9e-4ee6-9950-06ca90421e76","Type":"ContainerDied","Data":"fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613"} Dec 06 06:25:14 crc kubenswrapper[4957]: I1206 06:25:14.940658 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.022674 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/559c6b9f-c9c4-4248-aff8-79a7daff9442-kube-api-access-r59mp\") pod \"559c6b9f-c9c4-4248-aff8-79a7daff9442\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.023011 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-catalog-content\") pod \"559c6b9f-c9c4-4248-aff8-79a7daff9442\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.023151 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-utilities\") pod \"559c6b9f-c9c4-4248-aff8-79a7daff9442\" (UID: \"559c6b9f-c9c4-4248-aff8-79a7daff9442\") " Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.024856 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-utilities" (OuterVolumeSpecName: "utilities") pod "559c6b9f-c9c4-4248-aff8-79a7daff9442" (UID: "559c6b9f-c9c4-4248-aff8-79a7daff9442"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.030956 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/559c6b9f-c9c4-4248-aff8-79a7daff9442-kube-api-access-r59mp" (OuterVolumeSpecName: "kube-api-access-r59mp") pod "559c6b9f-c9c4-4248-aff8-79a7daff9442" (UID: "559c6b9f-c9c4-4248-aff8-79a7daff9442"). InnerVolumeSpecName "kube-api-access-r59mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.082884 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "559c6b9f-c9c4-4248-aff8-79a7daff9442" (UID: "559c6b9f-c9c4-4248-aff8-79a7daff9442"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.125694 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.125727 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/559c6b9f-c9c4-4248-aff8-79a7daff9442-kube-api-access-r59mp\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.125738 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559c6b9f-c9c4-4248-aff8-79a7daff9442-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.551865 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6kvh" event={"ID":"09ce914e-ef64-46dd-8aad-eafc79d08d2f","Type":"ContainerStarted","Data":"3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32"} Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.558467 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-khx5x" event={"ID":"559c6b9f-c9c4-4248-aff8-79a7daff9442","Type":"ContainerDied","Data":"53496ff3b42d06c41f9cf47ee4f45cc1e595679a2a1807aeb3bc10c51d01b2e8"} Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.558529 4957 scope.go:117] "RemoveContainer" containerID="3f7548394b9204d8bcaf9293f131877835153c08172a506c8f3917b4f51bc790" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.558690 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-khx5x" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.625554 4957 scope.go:117] "RemoveContainer" containerID="8b52aa6111a4455d2e8f4c7cc6fbb590268f06e02a38acf9f28f1e59b866b2cb" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.652064 4957 scope.go:117] "RemoveContainer" containerID="d13c1fa04295d8615bcf79373a2299a2a4ed78dfb7efa5bf2b5cb9fdb1bc4a06" Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.655174 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-khx5x"] Dec 06 06:25:15 crc kubenswrapper[4957]: I1206 06:25:15.664854 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-khx5x"] Dec 06 06:25:16 crc kubenswrapper[4957]: I1206 06:25:16.568758 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gp7zb" event={"ID":"a3f95673-ae9e-4ee6-9950-06ca90421e76","Type":"ContainerStarted","Data":"ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13"} Dec 06 06:25:16 crc kubenswrapper[4957]: I1206 06:25:16.571103 4957 generic.go:334] "Generic (PLEG): container finished" podID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerID="3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32" exitCode=0 Dec 06 06:25:16 crc kubenswrapper[4957]: I1206 06:25:16.571223 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6kvh" event={"ID":"09ce914e-ef64-46dd-8aad-eafc79d08d2f","Type":"ContainerDied","Data":"3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32"} Dec 06 06:25:16 crc kubenswrapper[4957]: I1206 06:25:16.673368 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" path="/var/lib/kubelet/pods/559c6b9f-c9c4-4248-aff8-79a7daff9442/volumes" Dec 06 06:25:19 crc kubenswrapper[4957]: I1206 06:25:19.601469 4957 generic.go:334] "Generic (PLEG): container finished" podID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerID="ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13" exitCode=0 Dec 06 06:25:19 crc kubenswrapper[4957]: I1206 06:25:19.601558 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gp7zb" event={"ID":"a3f95673-ae9e-4ee6-9950-06ca90421e76","Type":"ContainerDied","Data":"ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13"} Dec 06 06:25:21 crc kubenswrapper[4957]: I1206 06:25:21.625433 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gp7zb" event={"ID":"a3f95673-ae9e-4ee6-9950-06ca90421e76","Type":"ContainerStarted","Data":"d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6"} Dec 06 06:25:21 crc kubenswrapper[4957]: I1206 06:25:21.627786 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6kvh" event={"ID":"09ce914e-ef64-46dd-8aad-eafc79d08d2f","Type":"ContainerStarted","Data":"1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b"} Dec 06 06:25:21 crc kubenswrapper[4957]: I1206 06:25:21.657316 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gp7zb" podStartSLOduration=8.38693244 podStartE2EDuration="14.657297368s" podCreationTimestamp="2025-12-06 06:25:07 +0000 UTC" firstStartedPulling="2025-12-06 06:25:14.542063435 +0000 UTC m=+2754.192331087" lastFinishedPulling="2025-12-06 06:25:20.812428393 +0000 UTC m=+2760.462696015" observedRunningTime="2025-12-06 06:25:21.648502969 +0000 UTC m=+2761.298770621" watchObservedRunningTime="2025-12-06 06:25:21.657297368 +0000 UTC m=+2761.307565000" Dec 06 06:25:21 crc kubenswrapper[4957]: I1206 06:25:21.672741 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g6kvh" podStartSLOduration=5.749746198 podStartE2EDuration="11.672721207s" podCreationTimestamp="2025-12-06 06:25:10 +0000 UTC" firstStartedPulling="2025-12-06 06:25:14.533492432 +0000 UTC m=+2754.183760064" lastFinishedPulling="2025-12-06 06:25:20.456467421 +0000 UTC m=+2760.106735073" observedRunningTime="2025-12-06 06:25:21.666215791 +0000 UTC m=+2761.316483443" watchObservedRunningTime="2025-12-06 06:25:21.672721207 +0000 UTC m=+2761.322988829" Dec 06 06:25:27 crc kubenswrapper[4957]: I1206 06:25:27.805630 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:27 crc kubenswrapper[4957]: I1206 06:25:27.806376 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:27 crc kubenswrapper[4957]: I1206 06:25:27.850520 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:28 crc kubenswrapper[4957]: I1206 06:25:28.754172 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:28 crc kubenswrapper[4957]: I1206 06:25:28.807445 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gp7zb"] Dec 06 06:25:30 crc kubenswrapper[4957]: I1206 06:25:30.734781 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gp7zb" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="registry-server" containerID="cri-o://d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6" gracePeriod=2 Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.051084 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.051300 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.103804 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.209459 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.378461 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvcwd\" (UniqueName: \"kubernetes.io/projected/a3f95673-ae9e-4ee6-9950-06ca90421e76-kube-api-access-kvcwd\") pod \"a3f95673-ae9e-4ee6-9950-06ca90421e76\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.378525 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-catalog-content\") pod \"a3f95673-ae9e-4ee6-9950-06ca90421e76\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.378595 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-utilities\") pod \"a3f95673-ae9e-4ee6-9950-06ca90421e76\" (UID: \"a3f95673-ae9e-4ee6-9950-06ca90421e76\") " Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.379536 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-utilities" (OuterVolumeSpecName: "utilities") pod "a3f95673-ae9e-4ee6-9950-06ca90421e76" (UID: "a3f95673-ae9e-4ee6-9950-06ca90421e76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.385135 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f95673-ae9e-4ee6-9950-06ca90421e76-kube-api-access-kvcwd" (OuterVolumeSpecName: "kube-api-access-kvcwd") pod "a3f95673-ae9e-4ee6-9950-06ca90421e76" (UID: "a3f95673-ae9e-4ee6-9950-06ca90421e76"). InnerVolumeSpecName "kube-api-access-kvcwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.480875 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.480918 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvcwd\" (UniqueName: \"kubernetes.io/projected/a3f95673-ae9e-4ee6-9950-06ca90421e76-kube-api-access-kvcwd\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.482027 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3f95673-ae9e-4ee6-9950-06ca90421e76" (UID: "a3f95673-ae9e-4ee6-9950-06ca90421e76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.583090 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f95673-ae9e-4ee6-9950-06ca90421e76-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.745316 4957 generic.go:334] "Generic (PLEG): container finished" podID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerID="d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6" exitCode=0 Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.745384 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gp7zb" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.745403 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gp7zb" event={"ID":"a3f95673-ae9e-4ee6-9950-06ca90421e76","Type":"ContainerDied","Data":"d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6"} Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.746638 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gp7zb" event={"ID":"a3f95673-ae9e-4ee6-9950-06ca90421e76","Type":"ContainerDied","Data":"77e63140b6591559fe3342573ff9be39cc4b3a990627d41f25c32f2a21f6f4bd"} Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.746660 4957 scope.go:117] "RemoveContainer" containerID="d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.766844 4957 scope.go:117] "RemoveContainer" containerID="ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.778120 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gp7zb"] Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.788415 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gp7zb"] Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.807040 4957 scope.go:117] "RemoveContainer" containerID="fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.807104 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.838433 4957 scope.go:117] "RemoveContainer" containerID="d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6" Dec 06 06:25:31 crc kubenswrapper[4957]: E1206 06:25:31.839124 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6\": container with ID starting with d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6 not found: ID does not exist" containerID="d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.839188 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6"} err="failed to get container status \"d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6\": rpc error: code = NotFound desc = could not find container \"d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6\": container with ID starting with d69bfc343730dfafc02b72d510e5ed9a10d68fb414512495befc4178e552dfb6 not found: ID does not exist" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.839214 4957 scope.go:117] "RemoveContainer" containerID="ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13" Dec 06 06:25:31 crc kubenswrapper[4957]: E1206 06:25:31.839673 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13\": container with ID starting with ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13 not found: ID does not exist" containerID="ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.839717 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13"} err="failed to get container status \"ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13\": rpc error: code = NotFound desc = could not find container \"ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13\": container with ID starting with ec28b454d852b0e11670297ff4983ba5ba6d27b97948aebae89ff50d68e27d13 not found: ID does not exist" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.839744 4957 scope.go:117] "RemoveContainer" containerID="fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613" Dec 06 06:25:31 crc kubenswrapper[4957]: E1206 06:25:31.840071 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613\": container with ID starting with fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613 not found: ID does not exist" containerID="fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613" Dec 06 06:25:31 crc kubenswrapper[4957]: I1206 06:25:31.840130 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613"} err="failed to get container status \"fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613\": rpc error: code = NotFound desc = could not find container \"fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613\": container with ID starting with fb3edc8abb784a66086dba707b938725fb01ada9ff13fa43904b57694b28f613 not found: ID does not exist" Dec 06 06:25:32 crc kubenswrapper[4957]: I1206 06:25:32.672243 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" path="/var/lib/kubelet/pods/a3f95673-ae9e-4ee6-9950-06ca90421e76/volumes" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.085595 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6kvh"] Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.086164 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g6kvh" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="registry-server" containerID="cri-o://1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b" gracePeriod=2 Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.509102 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.640002 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75g5j\" (UniqueName: \"kubernetes.io/projected/09ce914e-ef64-46dd-8aad-eafc79d08d2f-kube-api-access-75g5j\") pod \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.640175 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-utilities\") pod \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.640321 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-catalog-content\") pod \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\" (UID: \"09ce914e-ef64-46dd-8aad-eafc79d08d2f\") " Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.641162 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-utilities" (OuterVolumeSpecName: "utilities") pod "09ce914e-ef64-46dd-8aad-eafc79d08d2f" (UID: "09ce914e-ef64-46dd-8aad-eafc79d08d2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.656169 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ce914e-ef64-46dd-8aad-eafc79d08d2f-kube-api-access-75g5j" (OuterVolumeSpecName: "kube-api-access-75g5j") pod "09ce914e-ef64-46dd-8aad-eafc79d08d2f" (UID: "09ce914e-ef64-46dd-8aad-eafc79d08d2f"). InnerVolumeSpecName "kube-api-access-75g5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.673695 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09ce914e-ef64-46dd-8aad-eafc79d08d2f" (UID: "09ce914e-ef64-46dd-8aad-eafc79d08d2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.742162 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.742205 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75g5j\" (UniqueName: \"kubernetes.io/projected/09ce914e-ef64-46dd-8aad-eafc79d08d2f-kube-api-access-75g5j\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.742218 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ce914e-ef64-46dd-8aad-eafc79d08d2f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.779161 4957 generic.go:334] "Generic (PLEG): container finished" podID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerID="1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b" exitCode=0 Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.779211 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6kvh" event={"ID":"09ce914e-ef64-46dd-8aad-eafc79d08d2f","Type":"ContainerDied","Data":"1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b"} Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.779242 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6kvh" event={"ID":"09ce914e-ef64-46dd-8aad-eafc79d08d2f","Type":"ContainerDied","Data":"333fe18f00772619bd0bbf21617b9908c552822a2129f1bc715bd2f64f21ffb2"} Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.779264 4957 scope.go:117] "RemoveContainer" containerID="1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.779270 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6kvh" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.811246 4957 scope.go:117] "RemoveContainer" containerID="3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.815065 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6kvh"] Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.836166 4957 scope.go:117] "RemoveContainer" containerID="b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.847278 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6kvh"] Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.874344 4957 scope.go:117] "RemoveContainer" containerID="1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b" Dec 06 06:25:34 crc kubenswrapper[4957]: E1206 06:25:34.874739 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b\": container with ID starting with 1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b not found: ID does not exist" containerID="1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.874780 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b"} err="failed to get container status \"1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b\": rpc error: code = NotFound desc = could not find container \"1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b\": container with ID starting with 1a597f48e6e7fda7d6fc2f613392d12500e9c3dd8a942af1c0a1ef4de1ba405b not found: ID does not exist" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.874810 4957 scope.go:117] "RemoveContainer" containerID="3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32" Dec 06 06:25:34 crc kubenswrapper[4957]: E1206 06:25:34.875213 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32\": container with ID starting with 3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32 not found: ID does not exist" containerID="3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.875241 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32"} err="failed to get container status \"3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32\": rpc error: code = NotFound desc = could not find container \"3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32\": container with ID starting with 3a4fa4d2d84b11dc77deec131110356c562ffaf008c0a7a8bd2d7262aa61fa32 not found: ID does not exist" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.875262 4957 scope.go:117] "RemoveContainer" containerID="b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084" Dec 06 06:25:34 crc kubenswrapper[4957]: E1206 06:25:34.875622 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084\": container with ID starting with b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084 not found: ID does not exist" containerID="b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084" Dec 06 06:25:34 crc kubenswrapper[4957]: I1206 06:25:34.875645 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084"} err="failed to get container status \"b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084\": rpc error: code = NotFound desc = could not find container \"b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084\": container with ID starting with b15e9624f5ad244f39cac0cc7f3f3274a98ae4292529ae8a9e3a7cad662b6084 not found: ID does not exist" Dec 06 06:25:36 crc kubenswrapper[4957]: I1206 06:25:36.674775 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" path="/var/lib/kubelet/pods/09ce914e-ef64-46dd-8aad-eafc79d08d2f/volumes" Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.181259 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.181673 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.181748 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.182611 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f17feb3bb43f81994797083be9593f7ff881e7b682b073303dbfac3dc781fc14"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.182686 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://f17feb3bb43f81994797083be9593f7ff881e7b682b073303dbfac3dc781fc14" gracePeriod=600 Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.845931 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="f17feb3bb43f81994797083be9593f7ff881e7b682b073303dbfac3dc781fc14" exitCode=0 Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.846011 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"f17feb3bb43f81994797083be9593f7ff881e7b682b073303dbfac3dc781fc14"} Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.846559 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536"} Dec 06 06:25:40 crc kubenswrapper[4957]: I1206 06:25:40.846590 4957 scope.go:117] "RemoveContainer" containerID="3ba910154c1389e418f272450a4ca5ffec5a7e59efe24ca40e59c439debb7092" Dec 06 06:25:42 crc kubenswrapper[4957]: I1206 06:25:42.874088 4957 generic.go:334] "Generic (PLEG): container finished" podID="5df54304-6d88-4027-ae21-40aaa328470f" containerID="e5acbcbe2dddfe1eb1024d16d744f1154a07a2717aa063088ce2b2ca63a1929a" exitCode=0 Dec 06 06:25:42 crc kubenswrapper[4957]: I1206 06:25:42.874707 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" event={"ID":"5df54304-6d88-4027-ae21-40aaa328470f","Type":"ContainerDied","Data":"e5acbcbe2dddfe1eb1024d16d744f1154a07a2717aa063088ce2b2ca63a1929a"} Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.356729 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.429884 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-combined-ca-bundle\") pod \"5df54304-6d88-4027-ae21-40aaa328470f\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.429940 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-inventory\") pod \"5df54304-6d88-4027-ae21-40aaa328470f\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.429964 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-ssh-key\") pod \"5df54304-6d88-4027-ae21-40aaa328470f\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.430067 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6fls\" (UniqueName: \"kubernetes.io/projected/5df54304-6d88-4027-ae21-40aaa328470f-kube-api-access-x6fls\") pod \"5df54304-6d88-4027-ae21-40aaa328470f\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.430090 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-secret-0\") pod \"5df54304-6d88-4027-ae21-40aaa328470f\" (UID: \"5df54304-6d88-4027-ae21-40aaa328470f\") " Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.435811 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5df54304-6d88-4027-ae21-40aaa328470f-kube-api-access-x6fls" (OuterVolumeSpecName: "kube-api-access-x6fls") pod "5df54304-6d88-4027-ae21-40aaa328470f" (UID: "5df54304-6d88-4027-ae21-40aaa328470f"). InnerVolumeSpecName "kube-api-access-x6fls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.436395 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5df54304-6d88-4027-ae21-40aaa328470f" (UID: "5df54304-6d88-4027-ae21-40aaa328470f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.458293 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-inventory" (OuterVolumeSpecName: "inventory") pod "5df54304-6d88-4027-ae21-40aaa328470f" (UID: "5df54304-6d88-4027-ae21-40aaa328470f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.460952 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5df54304-6d88-4027-ae21-40aaa328470f" (UID: "5df54304-6d88-4027-ae21-40aaa328470f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.468429 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "5df54304-6d88-4027-ae21-40aaa328470f" (UID: "5df54304-6d88-4027-ae21-40aaa328470f"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.532367 4957 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.532488 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.532508 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.532527 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6fls\" (UniqueName: \"kubernetes.io/projected/5df54304-6d88-4027-ae21-40aaa328470f-kube-api-access-x6fls\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.532546 4957 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5df54304-6d88-4027-ae21-40aaa328470f-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.891809 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" event={"ID":"5df54304-6d88-4027-ae21-40aaa328470f","Type":"ContainerDied","Data":"33838172f016da658ed6ec623e128ab0d2ad389e4376c83d9bb769cc5d9debea"} Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.892253 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33838172f016da658ed6ec623e128ab0d2ad389e4376c83d9bb769cc5d9debea" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.891878 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.993698 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69"] Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994195 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="extract-utilities" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994220 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="extract-utilities" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994243 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="extract-content" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994251 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="extract-content" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994261 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df54304-6d88-4027-ae21-40aaa328470f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994271 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df54304-6d88-4027-ae21-40aaa328470f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994287 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="extract-utilities" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994294 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="extract-utilities" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994306 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="extract-content" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994316 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="extract-content" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994331 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994338 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994352 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="extract-utilities" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994359 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="extract-utilities" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994373 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994381 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994397 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994405 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: E1206 06:25:44.994427 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="extract-content" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994434 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="extract-content" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994636 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f95673-ae9e-4ee6-9950-06ca90421e76" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994658 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="5df54304-6d88-4027-ae21-40aaa328470f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994705 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="559c6b9f-c9c4-4248-aff8-79a7daff9442" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.994719 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ce914e-ef64-46dd-8aad-eafc79d08d2f" containerName="registry-server" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.996165 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.998742 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.999393 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.999760 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.999798 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 06 06:25:44 crc kubenswrapper[4957]: I1206 06:25:44.999944 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.000750 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.005758 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.006288 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69"] Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.143244 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/312ec0f6-c580-4326-ba1a-b80706d87299-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.143311 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.143333 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.143372 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.144626 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.144679 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.144755 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.144791 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h2w9\" (UniqueName: \"kubernetes.io/projected/312ec0f6-c580-4326-ba1a-b80706d87299-kube-api-access-4h2w9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.144906 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246139 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246196 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/312ec0f6-c580-4326-ba1a-b80706d87299-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246247 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246270 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246317 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246383 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246407 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246468 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.246494 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h2w9\" (UniqueName: \"kubernetes.io/projected/312ec0f6-c580-4326-ba1a-b80706d87299-kube-api-access-4h2w9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.247478 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/312ec0f6-c580-4326-ba1a-b80706d87299-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.252592 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.253717 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.253717 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.254741 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.255215 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.256708 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.257370 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.266056 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h2w9\" (UniqueName: \"kubernetes.io/projected/312ec0f6-c580-4326-ba1a-b80706d87299-kube-api-access-4h2w9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dt69\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.320823 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.834457 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69"] Dec 06 06:25:45 crc kubenswrapper[4957]: I1206 06:25:45.901104 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" event={"ID":"312ec0f6-c580-4326-ba1a-b80706d87299","Type":"ContainerStarted","Data":"2743102db6c9abb39dbf69fc417456950717b94e09fb10ab909436d3880137cc"} Dec 06 06:25:46 crc kubenswrapper[4957]: I1206 06:25:46.911135 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" event={"ID":"312ec0f6-c580-4326-ba1a-b80706d87299","Type":"ContainerStarted","Data":"25101a018e45abbe7c475bf569ec92a33ae429b2deb4916b6fe5db54dbb599a9"} Dec 06 06:25:46 crc kubenswrapper[4957]: I1206 06:25:46.950167 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" podStartSLOduration=2.488541261 podStartE2EDuration="2.950139133s" podCreationTimestamp="2025-12-06 06:25:44 +0000 UTC" firstStartedPulling="2025-12-06 06:25:45.837089941 +0000 UTC m=+2785.487357583" lastFinishedPulling="2025-12-06 06:25:46.298687793 +0000 UTC m=+2785.948955455" observedRunningTime="2025-12-06 06:25:46.94337464 +0000 UTC m=+2786.593642282" watchObservedRunningTime="2025-12-06 06:25:46.950139133 +0000 UTC m=+2786.600406785" Dec 06 06:27:40 crc kubenswrapper[4957]: I1206 06:27:40.181532 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:27:40 crc kubenswrapper[4957]: I1206 06:27:40.182151 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:28:10 crc kubenswrapper[4957]: I1206 06:28:10.181286 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:28:10 crc kubenswrapper[4957]: I1206 06:28:10.182807 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:28:33 crc kubenswrapper[4957]: I1206 06:28:33.503892 4957 generic.go:334] "Generic (PLEG): container finished" podID="312ec0f6-c580-4326-ba1a-b80706d87299" containerID="25101a018e45abbe7c475bf569ec92a33ae429b2deb4916b6fe5db54dbb599a9" exitCode=0 Dec 06 06:28:33 crc kubenswrapper[4957]: I1206 06:28:33.504040 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" event={"ID":"312ec0f6-c580-4326-ba1a-b80706d87299","Type":"ContainerDied","Data":"25101a018e45abbe7c475bf569ec92a33ae429b2deb4916b6fe5db54dbb599a9"} Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.031800 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.097682 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-combined-ca-bundle\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.097766 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-inventory\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.097897 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-1\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.097940 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/312ec0f6-c580-4326-ba1a-b80706d87299-nova-extra-config-0\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.098002 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-ssh-key\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.098103 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-0\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.098221 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h2w9\" (UniqueName: \"kubernetes.io/projected/312ec0f6-c580-4326-ba1a-b80706d87299-kube-api-access-4h2w9\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.098281 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-0\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.098324 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-1\") pod \"312ec0f6-c580-4326-ba1a-b80706d87299\" (UID: \"312ec0f6-c580-4326-ba1a-b80706d87299\") " Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.118308 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/312ec0f6-c580-4326-ba1a-b80706d87299-kube-api-access-4h2w9" (OuterVolumeSpecName: "kube-api-access-4h2w9") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "kube-api-access-4h2w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.118345 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.124922 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/312ec0f6-c580-4326-ba1a-b80706d87299-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.128142 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.133513 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-inventory" (OuterVolumeSpecName: "inventory") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.133579 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.137987 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.146656 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.147590 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "312ec0f6-c580-4326-ba1a-b80706d87299" (UID: "312ec0f6-c580-4326-ba1a-b80706d87299"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200529 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h2w9\" (UniqueName: \"kubernetes.io/projected/312ec0f6-c580-4326-ba1a-b80706d87299-kube-api-access-4h2w9\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200579 4957 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200590 4957 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200598 4957 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200607 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200634 4957 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200644 4957 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/312ec0f6-c580-4326-ba1a-b80706d87299-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200669 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.200678 4957 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/312ec0f6-c580-4326-ba1a-b80706d87299-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.523975 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" event={"ID":"312ec0f6-c580-4326-ba1a-b80706d87299","Type":"ContainerDied","Data":"2743102db6c9abb39dbf69fc417456950717b94e09fb10ab909436d3880137cc"} Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.524325 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2743102db6c9abb39dbf69fc417456950717b94e09fb10ab909436d3880137cc" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.524048 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dt69" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.640895 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd"] Dec 06 06:28:35 crc kubenswrapper[4957]: E1206 06:28:35.641327 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="312ec0f6-c580-4326-ba1a-b80706d87299" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.641346 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="312ec0f6-c580-4326-ba1a-b80706d87299" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.641540 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="312ec0f6-c580-4326-ba1a-b80706d87299" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.642196 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.644559 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.644759 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.644797 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.645024 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.646261 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb789" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.652460 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd"] Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.710791 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.711070 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.712349 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzr7t\" (UniqueName: \"kubernetes.io/projected/37341858-eaac-4c32-a5cc-7b06843f4fc2-kube-api-access-tzr7t\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.712456 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.714562 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.714850 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.715153 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.816925 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.817003 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.817051 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.817079 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.817116 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.817160 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzr7t\" (UniqueName: \"kubernetes.io/projected/37341858-eaac-4c32-a5cc-7b06843f4fc2-kube-api-access-tzr7t\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.817193 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.821907 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.822262 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.822642 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.823159 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.823428 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.824462 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.834014 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzr7t\" (UniqueName: \"kubernetes.io/projected/37341858-eaac-4c32-a5cc-7b06843f4fc2-kube-api-access-tzr7t\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g78cd\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:35 crc kubenswrapper[4957]: I1206 06:28:35.960647 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:28:36 crc kubenswrapper[4957]: I1206 06:28:36.509687 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd"] Dec 06 06:28:36 crc kubenswrapper[4957]: I1206 06:28:36.534123 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" event={"ID":"37341858-eaac-4c32-a5cc-7b06843f4fc2","Type":"ContainerStarted","Data":"2417914fdd19b013169176fcec32c966adf8b613a1706a86f250a83e0b98a29b"} Dec 06 06:28:37 crc kubenswrapper[4957]: I1206 06:28:37.546711 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" event={"ID":"37341858-eaac-4c32-a5cc-7b06843f4fc2","Type":"ContainerStarted","Data":"a281c9cb1b2c83d767f222ceaf332a291510903ef87ef5ab4d4cc8cb567593e7"} Dec 06 06:28:37 crc kubenswrapper[4957]: I1206 06:28:37.576477 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" podStartSLOduration=2.171171313 podStartE2EDuration="2.576457595s" podCreationTimestamp="2025-12-06 06:28:35 +0000 UTC" firstStartedPulling="2025-12-06 06:28:36.518950572 +0000 UTC m=+2956.169218204" lastFinishedPulling="2025-12-06 06:28:36.924236854 +0000 UTC m=+2956.574504486" observedRunningTime="2025-12-06 06:28:37.567250015 +0000 UTC m=+2957.217517667" watchObservedRunningTime="2025-12-06 06:28:37.576457595 +0000 UTC m=+2957.226725227" Dec 06 06:28:40 crc kubenswrapper[4957]: I1206 06:28:40.181337 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:28:40 crc kubenswrapper[4957]: I1206 06:28:40.181680 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:28:40 crc kubenswrapper[4957]: I1206 06:28:40.181729 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:28:40 crc kubenswrapper[4957]: I1206 06:28:40.182528 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:28:40 crc kubenswrapper[4957]: I1206 06:28:40.182579 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" gracePeriod=600 Dec 06 06:28:42 crc kubenswrapper[4957]: E1206 06:28:42.130903 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:28:42 crc kubenswrapper[4957]: I1206 06:28:42.598430 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" exitCode=0 Dec 06 06:28:42 crc kubenswrapper[4957]: I1206 06:28:42.598556 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536"} Dec 06 06:28:42 crc kubenswrapper[4957]: I1206 06:28:42.598642 4957 scope.go:117] "RemoveContainer" containerID="f17feb3bb43f81994797083be9593f7ff881e7b682b073303dbfac3dc781fc14" Dec 06 06:28:42 crc kubenswrapper[4957]: I1206 06:28:42.599547 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:28:42 crc kubenswrapper[4957]: E1206 06:28:42.599850 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:28:55 crc kubenswrapper[4957]: I1206 06:28:55.667085 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:28:55 crc kubenswrapper[4957]: E1206 06:28:55.668052 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:29:08 crc kubenswrapper[4957]: I1206 06:29:08.662810 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:29:08 crc kubenswrapper[4957]: E1206 06:29:08.663828 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:29:20 crc kubenswrapper[4957]: I1206 06:29:20.669471 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:29:20 crc kubenswrapper[4957]: E1206 06:29:20.670302 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:29:33 crc kubenswrapper[4957]: I1206 06:29:33.662541 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:29:33 crc kubenswrapper[4957]: E1206 06:29:33.663368 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:29:49 crc kubenswrapper[4957]: I1206 06:29:49.508357 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:29:49 crc kubenswrapper[4957]: E1206 06:29:49.512093 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.164310 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp"] Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.166755 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.169207 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.169226 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.175869 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp"] Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.228756 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nww75\" (UniqueName: \"kubernetes.io/projected/4e054d25-40c3-4b50-8bf2-3fac94608b12-kube-api-access-nww75\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.229086 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e054d25-40c3-4b50-8bf2-3fac94608b12-secret-volume\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.229135 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e054d25-40c3-4b50-8bf2-3fac94608b12-config-volume\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.331165 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e054d25-40c3-4b50-8bf2-3fac94608b12-secret-volume\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.331207 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e054d25-40c3-4b50-8bf2-3fac94608b12-config-volume\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.331235 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nww75\" (UniqueName: \"kubernetes.io/projected/4e054d25-40c3-4b50-8bf2-3fac94608b12-kube-api-access-nww75\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.332346 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e054d25-40c3-4b50-8bf2-3fac94608b12-config-volume\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.337861 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e054d25-40c3-4b50-8bf2-3fac94608b12-secret-volume\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.348036 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nww75\" (UniqueName: \"kubernetes.io/projected/4e054d25-40c3-4b50-8bf2-3fac94608b12-kube-api-access-nww75\") pod \"collect-profiles-29416710-kxmdp\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.491745 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.671576 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:30:00 crc kubenswrapper[4957]: E1206 06:30:00.672225 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:30:00 crc kubenswrapper[4957]: I1206 06:30:00.927506 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp"] Dec 06 06:30:01 crc kubenswrapper[4957]: I1206 06:30:01.664353 4957 generic.go:334] "Generic (PLEG): container finished" podID="4e054d25-40c3-4b50-8bf2-3fac94608b12" containerID="4fd681aee97e1ebbdfd3fed896eee2d903f5bf807135d1eb2b5172399e4d6e8a" exitCode=0 Dec 06 06:30:01 crc kubenswrapper[4957]: I1206 06:30:01.664721 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" event={"ID":"4e054d25-40c3-4b50-8bf2-3fac94608b12","Type":"ContainerDied","Data":"4fd681aee97e1ebbdfd3fed896eee2d903f5bf807135d1eb2b5172399e4d6e8a"} Dec 06 06:30:01 crc kubenswrapper[4957]: I1206 06:30:01.664747 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" event={"ID":"4e054d25-40c3-4b50-8bf2-3fac94608b12","Type":"ContainerStarted","Data":"7dbd3fc09ef840737dddec2ec7a85e10f5f7baa9942dffe0fe8d09ec6e5bb16f"} Dec 06 06:30:02 crc kubenswrapper[4957]: I1206 06:30:02.995400 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.083739 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e054d25-40c3-4b50-8bf2-3fac94608b12-secret-volume\") pod \"4e054d25-40c3-4b50-8bf2-3fac94608b12\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.084028 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nww75\" (UniqueName: \"kubernetes.io/projected/4e054d25-40c3-4b50-8bf2-3fac94608b12-kube-api-access-nww75\") pod \"4e054d25-40c3-4b50-8bf2-3fac94608b12\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.085086 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e054d25-40c3-4b50-8bf2-3fac94608b12-config-volume\") pod \"4e054d25-40c3-4b50-8bf2-3fac94608b12\" (UID: \"4e054d25-40c3-4b50-8bf2-3fac94608b12\") " Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.085527 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e054d25-40c3-4b50-8bf2-3fac94608b12-config-volume" (OuterVolumeSpecName: "config-volume") pod "4e054d25-40c3-4b50-8bf2-3fac94608b12" (UID: "4e054d25-40c3-4b50-8bf2-3fac94608b12"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.090334 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e054d25-40c3-4b50-8bf2-3fac94608b12-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4e054d25-40c3-4b50-8bf2-3fac94608b12" (UID: "4e054d25-40c3-4b50-8bf2-3fac94608b12"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.090960 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e054d25-40c3-4b50-8bf2-3fac94608b12-kube-api-access-nww75" (OuterVolumeSpecName: "kube-api-access-nww75") pod "4e054d25-40c3-4b50-8bf2-3fac94608b12" (UID: "4e054d25-40c3-4b50-8bf2-3fac94608b12"). InnerVolumeSpecName "kube-api-access-nww75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.187779 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e054d25-40c3-4b50-8bf2-3fac94608b12-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.187824 4957 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e054d25-40c3-4b50-8bf2-3fac94608b12-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.187854 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nww75\" (UniqueName: \"kubernetes.io/projected/4e054d25-40c3-4b50-8bf2-3fac94608b12-kube-api-access-nww75\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.691536 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" event={"ID":"4e054d25-40c3-4b50-8bf2-3fac94608b12","Type":"ContainerDied","Data":"7dbd3fc09ef840737dddec2ec7a85e10f5f7baa9942dffe0fe8d09ec6e5bb16f"} Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.691944 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dbd3fc09ef840737dddec2ec7a85e10f5f7baa9942dffe0fe8d09ec6e5bb16f" Dec 06 06:30:03 crc kubenswrapper[4957]: I1206 06:30:03.691600 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416710-kxmdp" Dec 06 06:30:04 crc kubenswrapper[4957]: I1206 06:30:04.067878 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4"] Dec 06 06:30:04 crc kubenswrapper[4957]: I1206 06:30:04.076179 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416665-t95k4"] Dec 06 06:30:04 crc kubenswrapper[4957]: I1206 06:30:04.676458 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572c71f1-84ce-4b36-bd03-e17ee311d789" path="/var/lib/kubelet/pods/572c71f1-84ce-4b36-bd03-e17ee311d789/volumes" Dec 06 06:30:12 crc kubenswrapper[4957]: I1206 06:30:12.663118 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:30:12 crc kubenswrapper[4957]: E1206 06:30:12.664145 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:30:24 crc kubenswrapper[4957]: I1206 06:30:24.662884 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:30:24 crc kubenswrapper[4957]: E1206 06:30:24.663503 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:30:38 crc kubenswrapper[4957]: I1206 06:30:38.663791 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:30:38 crc kubenswrapper[4957]: E1206 06:30:38.665063 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:30:45 crc kubenswrapper[4957]: I1206 06:30:45.495817 4957 scope.go:117] "RemoveContainer" containerID="0df9fa0db471f32867dc2abed3c36956d65ab2c6238d4d76ddc9b930d74090c4" Dec 06 06:30:49 crc kubenswrapper[4957]: I1206 06:30:49.663400 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:30:49 crc kubenswrapper[4957]: E1206 06:30:49.664940 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:30:52 crc kubenswrapper[4957]: I1206 06:30:52.152709 4957 generic.go:334] "Generic (PLEG): container finished" podID="37341858-eaac-4c32-a5cc-7b06843f4fc2" containerID="a281c9cb1b2c83d767f222ceaf332a291510903ef87ef5ab4d4cc8cb567593e7" exitCode=0 Dec 06 06:30:52 crc kubenswrapper[4957]: I1206 06:30:52.152927 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" event={"ID":"37341858-eaac-4c32-a5cc-7b06843f4fc2","Type":"ContainerDied","Data":"a281c9cb1b2c83d767f222ceaf332a291510903ef87ef5ab4d4cc8cb567593e7"} Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.538070 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.696788 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-telemetry-combined-ca-bundle\") pod \"37341858-eaac-4c32-a5cc-7b06843f4fc2\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.697207 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-1\") pod \"37341858-eaac-4c32-a5cc-7b06843f4fc2\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.697240 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ssh-key\") pod \"37341858-eaac-4c32-a5cc-7b06843f4fc2\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.697279 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-0\") pod \"37341858-eaac-4c32-a5cc-7b06843f4fc2\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.697368 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-inventory\") pod \"37341858-eaac-4c32-a5cc-7b06843f4fc2\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.697505 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-2\") pod \"37341858-eaac-4c32-a5cc-7b06843f4fc2\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.697536 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzr7t\" (UniqueName: \"kubernetes.io/projected/37341858-eaac-4c32-a5cc-7b06843f4fc2-kube-api-access-tzr7t\") pod \"37341858-eaac-4c32-a5cc-7b06843f4fc2\" (UID: \"37341858-eaac-4c32-a5cc-7b06843f4fc2\") " Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.703141 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37341858-eaac-4c32-a5cc-7b06843f4fc2-kube-api-access-tzr7t" (OuterVolumeSpecName: "kube-api-access-tzr7t") pod "37341858-eaac-4c32-a5cc-7b06843f4fc2" (UID: "37341858-eaac-4c32-a5cc-7b06843f4fc2"). InnerVolumeSpecName "kube-api-access-tzr7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.703862 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "37341858-eaac-4c32-a5cc-7b06843f4fc2" (UID: "37341858-eaac-4c32-a5cc-7b06843f4fc2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.738289 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-inventory" (OuterVolumeSpecName: "inventory") pod "37341858-eaac-4c32-a5cc-7b06843f4fc2" (UID: "37341858-eaac-4c32-a5cc-7b06843f4fc2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.742337 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "37341858-eaac-4c32-a5cc-7b06843f4fc2" (UID: "37341858-eaac-4c32-a5cc-7b06843f4fc2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.747957 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "37341858-eaac-4c32-a5cc-7b06843f4fc2" (UID: "37341858-eaac-4c32-a5cc-7b06843f4fc2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.753878 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "37341858-eaac-4c32-a5cc-7b06843f4fc2" (UID: "37341858-eaac-4c32-a5cc-7b06843f4fc2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.755897 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "37341858-eaac-4c32-a5cc-7b06843f4fc2" (UID: "37341858-eaac-4c32-a5cc-7b06843f4fc2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.799937 4957 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.799969 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzr7t\" (UniqueName: \"kubernetes.io/projected/37341858-eaac-4c32-a5cc-7b06843f4fc2-kube-api-access-tzr7t\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.799981 4957 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.799990 4957 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.799999 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.800008 4957 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:53 crc kubenswrapper[4957]: I1206 06:30:53.800019 4957 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37341858-eaac-4c32-a5cc-7b06843f4fc2-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 06:30:54 crc kubenswrapper[4957]: I1206 06:30:54.172715 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" event={"ID":"37341858-eaac-4c32-a5cc-7b06843f4fc2","Type":"ContainerDied","Data":"2417914fdd19b013169176fcec32c966adf8b613a1706a86f250a83e0b98a29b"} Dec 06 06:30:54 crc kubenswrapper[4957]: I1206 06:30:54.172762 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2417914fdd19b013169176fcec32c966adf8b613a1706a86f250a83e0b98a29b" Dec 06 06:30:54 crc kubenswrapper[4957]: I1206 06:30:54.172826 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g78cd" Dec 06 06:31:01 crc kubenswrapper[4957]: I1206 06:31:01.663584 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:31:01 crc kubenswrapper[4957]: E1206 06:31:01.664751 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:31:14 crc kubenswrapper[4957]: I1206 06:31:14.662558 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:31:14 crc kubenswrapper[4957]: E1206 06:31:14.663536 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:31:27 crc kubenswrapper[4957]: I1206 06:31:27.662699 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:31:27 crc kubenswrapper[4957]: E1206 06:31:27.664195 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.030531 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wplr2"] Dec 06 06:31:37 crc kubenswrapper[4957]: E1206 06:31:37.031593 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37341858-eaac-4c32-a5cc-7b06843f4fc2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.031613 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="37341858-eaac-4c32-a5cc-7b06843f4fc2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 06:31:37 crc kubenswrapper[4957]: E1206 06:31:37.031647 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e054d25-40c3-4b50-8bf2-3fac94608b12" containerName="collect-profiles" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.031656 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e054d25-40c3-4b50-8bf2-3fac94608b12" containerName="collect-profiles" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.031925 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e054d25-40c3-4b50-8bf2-3fac94608b12" containerName="collect-profiles" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.031958 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="37341858-eaac-4c32-a5cc-7b06843f4fc2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.033628 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.070956 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wplr2"] Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.077280 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfrpr\" (UniqueName: \"kubernetes.io/projected/0c99e085-0b00-4783-980b-320f10d04c1f-kube-api-access-nfrpr\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.077330 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-utilities\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.077581 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-catalog-content\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.179108 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-catalog-content\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.179167 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfrpr\" (UniqueName: \"kubernetes.io/projected/0c99e085-0b00-4783-980b-320f10d04c1f-kube-api-access-nfrpr\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.179191 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-utilities\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.179791 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-catalog-content\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.179870 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-utilities\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.200410 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfrpr\" (UniqueName: \"kubernetes.io/projected/0c99e085-0b00-4783-980b-320f10d04c1f-kube-api-access-nfrpr\") pod \"certified-operators-wplr2\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.359006 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:37 crc kubenswrapper[4957]: I1206 06:31:37.876526 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wplr2"] Dec 06 06:31:38 crc kubenswrapper[4957]: E1206 06:31:38.261405 4957 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c99e085_0b00_4783_980b_320f10d04c1f.slice/crio-4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:31:38 crc kubenswrapper[4957]: I1206 06:31:38.584496 4957 generic.go:334] "Generic (PLEG): container finished" podID="0c99e085-0b00-4783-980b-320f10d04c1f" containerID="4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7" exitCode=0 Dec 06 06:31:38 crc kubenswrapper[4957]: I1206 06:31:38.584549 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wplr2" event={"ID":"0c99e085-0b00-4783-980b-320f10d04c1f","Type":"ContainerDied","Data":"4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7"} Dec 06 06:31:38 crc kubenswrapper[4957]: I1206 06:31:38.584927 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wplr2" event={"ID":"0c99e085-0b00-4783-980b-320f10d04c1f","Type":"ContainerStarted","Data":"1d68d67c603994ac081da434c405af09e17e067ed21ccee18f3ccffeb1022f55"} Dec 06 06:31:38 crc kubenswrapper[4957]: I1206 06:31:38.587195 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:31:38 crc kubenswrapper[4957]: I1206 06:31:38.663266 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:31:38 crc kubenswrapper[4957]: E1206 06:31:38.663600 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:31:40 crc kubenswrapper[4957]: I1206 06:31:40.603032 4957 generic.go:334] "Generic (PLEG): container finished" podID="0c99e085-0b00-4783-980b-320f10d04c1f" containerID="bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5" exitCode=0 Dec 06 06:31:40 crc kubenswrapper[4957]: I1206 06:31:40.603632 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wplr2" event={"ID":"0c99e085-0b00-4783-980b-320f10d04c1f","Type":"ContainerDied","Data":"bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5"} Dec 06 06:31:41 crc kubenswrapper[4957]: I1206 06:31:41.615461 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wplr2" event={"ID":"0c99e085-0b00-4783-980b-320f10d04c1f","Type":"ContainerStarted","Data":"b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c"} Dec 06 06:31:41 crc kubenswrapper[4957]: I1206 06:31:41.649384 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wplr2" podStartSLOduration=2.241212518 podStartE2EDuration="4.649363623s" podCreationTimestamp="2025-12-06 06:31:37 +0000 UTC" firstStartedPulling="2025-12-06 06:31:38.586728645 +0000 UTC m=+3138.236996297" lastFinishedPulling="2025-12-06 06:31:40.99487977 +0000 UTC m=+3140.645147402" observedRunningTime="2025-12-06 06:31:41.639513545 +0000 UTC m=+3141.289781187" watchObservedRunningTime="2025-12-06 06:31:41.649363623 +0000 UTC m=+3141.299631265" Dec 06 06:31:47 crc kubenswrapper[4957]: I1206 06:31:47.360187 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:47 crc kubenswrapper[4957]: I1206 06:31:47.360822 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:47 crc kubenswrapper[4957]: I1206 06:31:47.456074 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:47 crc kubenswrapper[4957]: I1206 06:31:47.709826 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:47 crc kubenswrapper[4957]: I1206 06:31:47.763116 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wplr2"] Dec 06 06:31:49 crc kubenswrapper[4957]: I1206 06:31:49.678974 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wplr2" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="registry-server" containerID="cri-o://b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c" gracePeriod=2 Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.192293 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.341288 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-catalog-content\") pod \"0c99e085-0b00-4783-980b-320f10d04c1f\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.341481 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfrpr\" (UniqueName: \"kubernetes.io/projected/0c99e085-0b00-4783-980b-320f10d04c1f-kube-api-access-nfrpr\") pod \"0c99e085-0b00-4783-980b-320f10d04c1f\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.341623 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-utilities\") pod \"0c99e085-0b00-4783-980b-320f10d04c1f\" (UID: \"0c99e085-0b00-4783-980b-320f10d04c1f\") " Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.342431 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-utilities" (OuterVolumeSpecName: "utilities") pod "0c99e085-0b00-4783-980b-320f10d04c1f" (UID: "0c99e085-0b00-4783-980b-320f10d04c1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.342984 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.347347 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c99e085-0b00-4783-980b-320f10d04c1f-kube-api-access-nfrpr" (OuterVolumeSpecName: "kube-api-access-nfrpr") pod "0c99e085-0b00-4783-980b-320f10d04c1f" (UID: "0c99e085-0b00-4783-980b-320f10d04c1f"). InnerVolumeSpecName "kube-api-access-nfrpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.445479 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfrpr\" (UniqueName: \"kubernetes.io/projected/0c99e085-0b00-4783-980b-320f10d04c1f-kube-api-access-nfrpr\") on node \"crc\" DevicePath \"\"" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.634638 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 06:31:50 crc kubenswrapper[4957]: E1206 06:31:50.635152 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="registry-server" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.635174 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="registry-server" Dec 06 06:31:50 crc kubenswrapper[4957]: E1206 06:31:50.635218 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="extract-content" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.635226 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="extract-content" Dec 06 06:31:50 crc kubenswrapper[4957]: E1206 06:31:50.635244 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="extract-utilities" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.635252 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="extract-utilities" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.635504 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" containerName="registry-server" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.636359 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.638697 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-m7bqq" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.640119 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.640511 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.641422 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.644544 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.655037 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c99e085-0b00-4783-980b-320f10d04c1f" (UID: "0c99e085-0b00-4783-980b-320f10d04c1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.693025 4957 generic.go:334] "Generic (PLEG): container finished" podID="0c99e085-0b00-4783-980b-320f10d04c1f" containerID="b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c" exitCode=0 Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.693067 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wplr2" event={"ID":"0c99e085-0b00-4783-980b-320f10d04c1f","Type":"ContainerDied","Data":"b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c"} Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.693080 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wplr2" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.693092 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wplr2" event={"ID":"0c99e085-0b00-4783-980b-320f10d04c1f","Type":"ContainerDied","Data":"1d68d67c603994ac081da434c405af09e17e067ed21ccee18f3ccffeb1022f55"} Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.693138 4957 scope.go:117] "RemoveContainer" containerID="b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.722450 4957 scope.go:117] "RemoveContainer" containerID="bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.726035 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wplr2"] Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.737291 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wplr2"] Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754569 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754621 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754645 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754670 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754695 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754727 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-config-data\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754761 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754779 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754795 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd7qq\" (UniqueName: \"kubernetes.io/projected/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-kube-api-access-wd7qq\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.754873 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c99e085-0b00-4783-980b-320f10d04c1f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.770158 4957 scope.go:117] "RemoveContainer" containerID="4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.795819 4957 scope.go:117] "RemoveContainer" containerID="b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c" Dec 06 06:31:50 crc kubenswrapper[4957]: E1206 06:31:50.796332 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c\": container with ID starting with b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c not found: ID does not exist" containerID="b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.796366 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c"} err="failed to get container status \"b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c\": rpc error: code = NotFound desc = could not find container \"b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c\": container with ID starting with b2b1811964a0edd9ea10add0041c01402197ed5a174219d49a089e59ef1b468c not found: ID does not exist" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.796388 4957 scope.go:117] "RemoveContainer" containerID="bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5" Dec 06 06:31:50 crc kubenswrapper[4957]: E1206 06:31:50.796716 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5\": container with ID starting with bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5 not found: ID does not exist" containerID="bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.796735 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5"} err="failed to get container status \"bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5\": rpc error: code = NotFound desc = could not find container \"bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5\": container with ID starting with bfc1f518fde4c74e6a714c63dfd8e807f4eceea6610b65cd32c772ca527b85d5 not found: ID does not exist" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.796747 4957 scope.go:117] "RemoveContainer" containerID="4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7" Dec 06 06:31:50 crc kubenswrapper[4957]: E1206 06:31:50.797081 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7\": container with ID starting with 4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7 not found: ID does not exist" containerID="4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.797114 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7"} err="failed to get container status \"4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7\": rpc error: code = NotFound desc = could not find container \"4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7\": container with ID starting with 4d97bcda305c3055cfc309a21a09b216705a4a7607183a792662689fca0c56e7 not found: ID does not exist" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.856675 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.856750 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.856779 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.856817 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.856875 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.856933 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-config-data\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.856987 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.857011 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.857033 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd7qq\" (UniqueName: \"kubernetes.io/projected/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-kube-api-access-wd7qq\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.857392 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.858192 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.858255 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.859074 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-config-data\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.859114 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.862155 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.862874 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.868027 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.881715 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd7qq\" (UniqueName: \"kubernetes.io/projected/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-kube-api-access-wd7qq\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.897686 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " pod="openstack/tempest-tests-tempest" Dec 06 06:31:50 crc kubenswrapper[4957]: I1206 06:31:50.959211 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 06:31:51 crc kubenswrapper[4957]: I1206 06:31:51.392654 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 06:31:51 crc kubenswrapper[4957]: I1206 06:31:51.662994 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:31:51 crc kubenswrapper[4957]: E1206 06:31:51.663305 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:31:51 crc kubenswrapper[4957]: I1206 06:31:51.704295 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd","Type":"ContainerStarted","Data":"abc796ec6778cae473e15245a2df316c5f6481db47a112062264f412314de8da"} Dec 06 06:31:52 crc kubenswrapper[4957]: I1206 06:31:52.673419 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c99e085-0b00-4783-980b-320f10d04c1f" path="/var/lib/kubelet/pods/0c99e085-0b00-4783-980b-320f10d04c1f/volumes" Dec 06 06:32:02 crc kubenswrapper[4957]: I1206 06:32:02.663520 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:32:02 crc kubenswrapper[4957]: E1206 06:32:02.665062 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:32:17 crc kubenswrapper[4957]: I1206 06:32:17.662129 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:32:17 crc kubenswrapper[4957]: E1206 06:32:17.663081 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:32:26 crc kubenswrapper[4957]: E1206 06:32:26.232026 4957 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 06 06:32:26 crc kubenswrapper[4957]: E1206 06:32:26.233037 4957 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wd7qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(469a494f-1ceb-4d9f-bfe5-e0856e88f5bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 06:32:26 crc kubenswrapper[4957]: E1206 06:32:26.234263 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" Dec 06 06:32:27 crc kubenswrapper[4957]: E1206 06:32:27.045302 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" Dec 06 06:32:32 crc kubenswrapper[4957]: I1206 06:32:32.664635 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:32:32 crc kubenswrapper[4957]: E1206 06:32:32.665635 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:32:38 crc kubenswrapper[4957]: I1206 06:32:38.156868 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 06:32:39 crc kubenswrapper[4957]: I1206 06:32:39.141817 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd","Type":"ContainerStarted","Data":"2d93e442b77f6a1c41b471530fa69a446fe5b5b61d3c33320d7876e25efc2f5c"} Dec 06 06:32:47 crc kubenswrapper[4957]: I1206 06:32:47.662165 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:32:47 crc kubenswrapper[4957]: E1206 06:32:47.662977 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:33:00 crc kubenswrapper[4957]: I1206 06:33:00.689639 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:33:00 crc kubenswrapper[4957]: E1206 06:33:00.690421 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:33:15 crc kubenswrapper[4957]: I1206 06:33:15.663669 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:33:15 crc kubenswrapper[4957]: E1206 06:33:15.664537 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:33:30 crc kubenswrapper[4957]: I1206 06:33:30.668199 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:33:30 crc kubenswrapper[4957]: E1206 06:33:30.668967 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:33:42 crc kubenswrapper[4957]: I1206 06:33:42.662148 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:33:43 crc kubenswrapper[4957]: I1206 06:33:43.824869 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"d4e3af660d7ea6175b7239e78bf6b4827a27406240b47d1ba1e1552fb3eca408"} Dec 06 06:33:43 crc kubenswrapper[4957]: I1206 06:33:43.857191 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=68.098972648 podStartE2EDuration="1m54.857169152s" podCreationTimestamp="2025-12-06 06:31:49 +0000 UTC" firstStartedPulling="2025-12-06 06:31:51.395957942 +0000 UTC m=+3151.046225594" lastFinishedPulling="2025-12-06 06:32:38.154154466 +0000 UTC m=+3197.804422098" observedRunningTime="2025-12-06 06:32:39.172927308 +0000 UTC m=+3198.823194940" watchObservedRunningTime="2025-12-06 06:33:43.857169152 +0000 UTC m=+3263.507436784" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.600171 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jwl25"] Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.603764 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.621582 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jwl25"] Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.694911 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-utilities\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.695093 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-catalog-content\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.695121 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2df7q\" (UniqueName: \"kubernetes.io/projected/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-kube-api-access-2df7q\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.797285 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-utilities\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.797412 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-catalog-content\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.797432 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2df7q\" (UniqueName: \"kubernetes.io/projected/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-kube-api-access-2df7q\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.798784 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-utilities\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.799044 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-catalog-content\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.816375 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2df7q\" (UniqueName: \"kubernetes.io/projected/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-kube-api-access-2df7q\") pod \"redhat-operators-jwl25\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:37 crc kubenswrapper[4957]: I1206 06:35:37.965930 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:38 crc kubenswrapper[4957]: I1206 06:35:38.433576 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jwl25"] Dec 06 06:35:38 crc kubenswrapper[4957]: I1206 06:35:38.879462 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwl25" event={"ID":"07be19bd-56c3-4c8b-86af-b1a1ef3d757e","Type":"ContainerStarted","Data":"adf26f5e9d2c70d3dcc82cf460bc3aec38f1fbe0f42f2059861d2c2108864340"} Dec 06 06:35:39 crc kubenswrapper[4957]: I1206 06:35:39.889576 4957 generic.go:334] "Generic (PLEG): container finished" podID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerID="e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719" exitCode=0 Dec 06 06:35:39 crc kubenswrapper[4957]: I1206 06:35:39.889617 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwl25" event={"ID":"07be19bd-56c3-4c8b-86af-b1a1ef3d757e","Type":"ContainerDied","Data":"e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719"} Dec 06 06:35:40 crc kubenswrapper[4957]: I1206 06:35:40.899441 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwl25" event={"ID":"07be19bd-56c3-4c8b-86af-b1a1ef3d757e","Type":"ContainerStarted","Data":"61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd"} Dec 06 06:35:42 crc kubenswrapper[4957]: I1206 06:35:42.933373 4957 generic.go:334] "Generic (PLEG): container finished" podID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerID="61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd" exitCode=0 Dec 06 06:35:42 crc kubenswrapper[4957]: I1206 06:35:42.933503 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwl25" event={"ID":"07be19bd-56c3-4c8b-86af-b1a1ef3d757e","Type":"ContainerDied","Data":"61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd"} Dec 06 06:35:43 crc kubenswrapper[4957]: I1206 06:35:43.948429 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwl25" event={"ID":"07be19bd-56c3-4c8b-86af-b1a1ef3d757e","Type":"ContainerStarted","Data":"ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd"} Dec 06 06:35:43 crc kubenswrapper[4957]: I1206 06:35:43.990232 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jwl25" podStartSLOduration=3.563672198 podStartE2EDuration="6.990207981s" podCreationTimestamp="2025-12-06 06:35:37 +0000 UTC" firstStartedPulling="2025-12-06 06:35:39.891494474 +0000 UTC m=+3379.541762106" lastFinishedPulling="2025-12-06 06:35:43.318030257 +0000 UTC m=+3382.968297889" observedRunningTime="2025-12-06 06:35:43.983007376 +0000 UTC m=+3383.633275058" watchObservedRunningTime="2025-12-06 06:35:43.990207981 +0000 UTC m=+3383.640475623" Dec 06 06:35:47 crc kubenswrapper[4957]: I1206 06:35:47.967607 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:47 crc kubenswrapper[4957]: I1206 06:35:47.968317 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:49 crc kubenswrapper[4957]: I1206 06:35:49.032622 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jwl25" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="registry-server" probeResult="failure" output=< Dec 06 06:35:49 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 06:35:49 crc kubenswrapper[4957]: > Dec 06 06:35:58 crc kubenswrapper[4957]: I1206 06:35:58.018102 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:58 crc kubenswrapper[4957]: I1206 06:35:58.064725 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:58 crc kubenswrapper[4957]: I1206 06:35:58.257710 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jwl25"] Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.082174 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jwl25" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="registry-server" containerID="cri-o://ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd" gracePeriod=2 Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.567575 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.649318 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-utilities\") pod \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.649512 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-catalog-content\") pod \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.649652 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2df7q\" (UniqueName: \"kubernetes.io/projected/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-kube-api-access-2df7q\") pod \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\" (UID: \"07be19bd-56c3-4c8b-86af-b1a1ef3d757e\") " Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.650282 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-utilities" (OuterVolumeSpecName: "utilities") pod "07be19bd-56c3-4c8b-86af-b1a1ef3d757e" (UID: "07be19bd-56c3-4c8b-86af-b1a1ef3d757e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.655003 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-kube-api-access-2df7q" (OuterVolumeSpecName: "kube-api-access-2df7q") pod "07be19bd-56c3-4c8b-86af-b1a1ef3d757e" (UID: "07be19bd-56c3-4c8b-86af-b1a1ef3d757e"). InnerVolumeSpecName "kube-api-access-2df7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.752616 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.752669 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2df7q\" (UniqueName: \"kubernetes.io/projected/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-kube-api-access-2df7q\") on node \"crc\" DevicePath \"\"" Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.763264 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07be19bd-56c3-4c8b-86af-b1a1ef3d757e" (UID: "07be19bd-56c3-4c8b-86af-b1a1ef3d757e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:35:59 crc kubenswrapper[4957]: I1206 06:35:59.854272 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07be19bd-56c3-4c8b-86af-b1a1ef3d757e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.093711 4957 generic.go:334] "Generic (PLEG): container finished" podID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerID="ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd" exitCode=0 Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.093757 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwl25" event={"ID":"07be19bd-56c3-4c8b-86af-b1a1ef3d757e","Type":"ContainerDied","Data":"ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd"} Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.093789 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwl25" event={"ID":"07be19bd-56c3-4c8b-86af-b1a1ef3d757e","Type":"ContainerDied","Data":"adf26f5e9d2c70d3dcc82cf460bc3aec38f1fbe0f42f2059861d2c2108864340"} Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.093807 4957 scope.go:117] "RemoveContainer" containerID="ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.094899 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwl25" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.120483 4957 scope.go:117] "RemoveContainer" containerID="61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.143734 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jwl25"] Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.156042 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jwl25"] Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.168264 4957 scope.go:117] "RemoveContainer" containerID="e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.196010 4957 scope.go:117] "RemoveContainer" containerID="ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd" Dec 06 06:36:00 crc kubenswrapper[4957]: E1206 06:36:00.197151 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd\": container with ID starting with ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd not found: ID does not exist" containerID="ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.197183 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd"} err="failed to get container status \"ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd\": rpc error: code = NotFound desc = could not find container \"ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd\": container with ID starting with ef89208d88853275fac4324693f421f7f6f84c42971660be06ca9f9a2f161bcd not found: ID does not exist" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.197205 4957 scope.go:117] "RemoveContainer" containerID="61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd" Dec 06 06:36:00 crc kubenswrapper[4957]: E1206 06:36:00.197391 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd\": container with ID starting with 61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd not found: ID does not exist" containerID="61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.197411 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd"} err="failed to get container status \"61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd\": rpc error: code = NotFound desc = could not find container \"61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd\": container with ID starting with 61877ecc41ddc9eb7f705ef85ea5eeb4bb9d9e76577dcf6121a13c4a411843cd not found: ID does not exist" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.197426 4957 scope.go:117] "RemoveContainer" containerID="e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719" Dec 06 06:36:00 crc kubenswrapper[4957]: E1206 06:36:00.197688 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719\": container with ID starting with e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719 not found: ID does not exist" containerID="e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.197707 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719"} err="failed to get container status \"e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719\": rpc error: code = NotFound desc = could not find container \"e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719\": container with ID starting with e2d630581cab0266972dae0f7682aaa7d10b6e002db6d71b602379c749888719 not found: ID does not exist" Dec 06 06:36:00 crc kubenswrapper[4957]: I1206 06:36:00.674537 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" path="/var/lib/kubelet/pods/07be19bd-56c3-4c8b-86af-b1a1ef3d757e/volumes" Dec 06 06:36:10 crc kubenswrapper[4957]: I1206 06:36:10.181182 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:36:10 crc kubenswrapper[4957]: I1206 06:36:10.181764 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:36:14 crc kubenswrapper[4957]: I1206 06:36:14.955231 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wd9"] Dec 06 06:36:14 crc kubenswrapper[4957]: E1206 06:36:14.956303 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="extract-content" Dec 06 06:36:14 crc kubenswrapper[4957]: I1206 06:36:14.956318 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="extract-content" Dec 06 06:36:14 crc kubenswrapper[4957]: E1206 06:36:14.956357 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="registry-server" Dec 06 06:36:14 crc kubenswrapper[4957]: I1206 06:36:14.956363 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="registry-server" Dec 06 06:36:14 crc kubenswrapper[4957]: E1206 06:36:14.956383 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="extract-utilities" Dec 06 06:36:14 crc kubenswrapper[4957]: I1206 06:36:14.956392 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="extract-utilities" Dec 06 06:36:14 crc kubenswrapper[4957]: I1206 06:36:14.956641 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="07be19bd-56c3-4c8b-86af-b1a1ef3d757e" containerName="registry-server" Dec 06 06:36:14 crc kubenswrapper[4957]: I1206 06:36:14.958288 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:14 crc kubenswrapper[4957]: I1206 06:36:14.971482 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wd9"] Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.037981 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klf5s\" (UniqueName: \"kubernetes.io/projected/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-kube-api-access-klf5s\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.038042 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-catalog-content\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.038142 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-utilities\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.139236 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klf5s\" (UniqueName: \"kubernetes.io/projected/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-kube-api-access-klf5s\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.139297 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-catalog-content\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.139365 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-utilities\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.139855 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-catalog-content\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.141546 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-utilities\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.158569 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klf5s\" (UniqueName: \"kubernetes.io/projected/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-kube-api-access-klf5s\") pod \"redhat-marketplace-f4wd9\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.300456 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:15 crc kubenswrapper[4957]: I1206 06:36:15.779527 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wd9"] Dec 06 06:36:16 crc kubenswrapper[4957]: I1206 06:36:16.236232 4957 generic.go:334] "Generic (PLEG): container finished" podID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerID="d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b" exitCode=0 Dec 06 06:36:16 crc kubenswrapper[4957]: I1206 06:36:16.236300 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wd9" event={"ID":"f5426bc1-cb43-4c49-b2e5-b8ac43855c29","Type":"ContainerDied","Data":"d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b"} Dec 06 06:36:16 crc kubenswrapper[4957]: I1206 06:36:16.236509 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wd9" event={"ID":"f5426bc1-cb43-4c49-b2e5-b8ac43855c29","Type":"ContainerStarted","Data":"df3531886833f888c1a5f4368763be7771abb8544ec16ba18ae88bbfcdfa2705"} Dec 06 06:36:17 crc kubenswrapper[4957]: I1206 06:36:17.249028 4957 generic.go:334] "Generic (PLEG): container finished" podID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerID="ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d" exitCode=0 Dec 06 06:36:17 crc kubenswrapper[4957]: I1206 06:36:17.249365 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wd9" event={"ID":"f5426bc1-cb43-4c49-b2e5-b8ac43855c29","Type":"ContainerDied","Data":"ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d"} Dec 06 06:36:18 crc kubenswrapper[4957]: I1206 06:36:18.262680 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wd9" event={"ID":"f5426bc1-cb43-4c49-b2e5-b8ac43855c29","Type":"ContainerStarted","Data":"517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f"} Dec 06 06:36:18 crc kubenswrapper[4957]: I1206 06:36:18.283764 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f4wd9" podStartSLOduration=2.645682904 podStartE2EDuration="4.283748796s" podCreationTimestamp="2025-12-06 06:36:14 +0000 UTC" firstStartedPulling="2025-12-06 06:36:16.238015051 +0000 UTC m=+3415.888282683" lastFinishedPulling="2025-12-06 06:36:17.876080933 +0000 UTC m=+3417.526348575" observedRunningTime="2025-12-06 06:36:18.278783661 +0000 UTC m=+3417.929051293" watchObservedRunningTime="2025-12-06 06:36:18.283748796 +0000 UTC m=+3417.934016428" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.331237 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tts4c"] Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.334402 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.342422 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tts4c"] Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.421991 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-catalog-content\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.422396 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-utilities\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.422510 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84z8m\" (UniqueName: \"kubernetes.io/projected/2903e894-9c88-41e4-abf4-9532e09cd605-kube-api-access-84z8m\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.524218 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-utilities\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.524280 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84z8m\" (UniqueName: \"kubernetes.io/projected/2903e894-9c88-41e4-abf4-9532e09cd605-kube-api-access-84z8m\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.524355 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-catalog-content\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.524927 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-utilities\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.524966 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-catalog-content\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.545792 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84z8m\" (UniqueName: \"kubernetes.io/projected/2903e894-9c88-41e4-abf4-9532e09cd605-kube-api-access-84z8m\") pod \"community-operators-tts4c\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:24 crc kubenswrapper[4957]: I1206 06:36:24.655123 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:25 crc kubenswrapper[4957]: I1206 06:36:25.214987 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tts4c"] Dec 06 06:36:25 crc kubenswrapper[4957]: I1206 06:36:25.300618 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:25 crc kubenswrapper[4957]: I1206 06:36:25.300665 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:25 crc kubenswrapper[4957]: I1206 06:36:25.333236 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tts4c" event={"ID":"2903e894-9c88-41e4-abf4-9532e09cd605","Type":"ContainerStarted","Data":"e1b06a40c56fb52c84e76c343ce0d3118e416d380e72d551bca53e897082647f"} Dec 06 06:36:25 crc kubenswrapper[4957]: I1206 06:36:25.350591 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:25 crc kubenswrapper[4957]: I1206 06:36:25.410442 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:26 crc kubenswrapper[4957]: I1206 06:36:26.348805 4957 generic.go:334] "Generic (PLEG): container finished" podID="2903e894-9c88-41e4-abf4-9532e09cd605" containerID="af07479233ddd263e854288692111be7b407b29c41501df8f40697f1e25f7b88" exitCode=0 Dec 06 06:36:26 crc kubenswrapper[4957]: I1206 06:36:26.348872 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tts4c" event={"ID":"2903e894-9c88-41e4-abf4-9532e09cd605","Type":"ContainerDied","Data":"af07479233ddd263e854288692111be7b407b29c41501df8f40697f1e25f7b88"} Dec 06 06:36:27 crc kubenswrapper[4957]: I1206 06:36:27.360024 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tts4c" event={"ID":"2903e894-9c88-41e4-abf4-9532e09cd605","Type":"ContainerStarted","Data":"bfd8d5749eb88035a3f0d9ef6708f0128cb886505611d27f1cce26e6820ba0ff"} Dec 06 06:36:27 crc kubenswrapper[4957]: I1206 06:36:27.699375 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wd9"] Dec 06 06:36:27 crc kubenswrapper[4957]: I1206 06:36:27.699651 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f4wd9" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="registry-server" containerID="cri-o://517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f" gracePeriod=2 Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.296612 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.372112 4957 generic.go:334] "Generic (PLEG): container finished" podID="2903e894-9c88-41e4-abf4-9532e09cd605" containerID="bfd8d5749eb88035a3f0d9ef6708f0128cb886505611d27f1cce26e6820ba0ff" exitCode=0 Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.372218 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tts4c" event={"ID":"2903e894-9c88-41e4-abf4-9532e09cd605","Type":"ContainerDied","Data":"bfd8d5749eb88035a3f0d9ef6708f0128cb886505611d27f1cce26e6820ba0ff"} Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.375341 4957 generic.go:334] "Generic (PLEG): container finished" podID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerID="517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f" exitCode=0 Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.375419 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wd9" event={"ID":"f5426bc1-cb43-4c49-b2e5-b8ac43855c29","Type":"ContainerDied","Data":"517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f"} Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.375453 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f4wd9" event={"ID":"f5426bc1-cb43-4c49-b2e5-b8ac43855c29","Type":"ContainerDied","Data":"df3531886833f888c1a5f4368763be7771abb8544ec16ba18ae88bbfcdfa2705"} Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.375473 4957 scope.go:117] "RemoveContainer" containerID="517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.375602 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f4wd9" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.400621 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-utilities\") pod \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.400717 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-catalog-content\") pod \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.400949 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klf5s\" (UniqueName: \"kubernetes.io/projected/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-kube-api-access-klf5s\") pod \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\" (UID: \"f5426bc1-cb43-4c49-b2e5-b8ac43855c29\") " Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.403972 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-utilities" (OuterVolumeSpecName: "utilities") pod "f5426bc1-cb43-4c49-b2e5-b8ac43855c29" (UID: "f5426bc1-cb43-4c49-b2e5-b8ac43855c29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.408019 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-kube-api-access-klf5s" (OuterVolumeSpecName: "kube-api-access-klf5s") pod "f5426bc1-cb43-4c49-b2e5-b8ac43855c29" (UID: "f5426bc1-cb43-4c49-b2e5-b8ac43855c29"). InnerVolumeSpecName "kube-api-access-klf5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.415772 4957 scope.go:117] "RemoveContainer" containerID="ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.425162 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5426bc1-cb43-4c49-b2e5-b8ac43855c29" (UID: "f5426bc1-cb43-4c49-b2e5-b8ac43855c29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.463655 4957 scope.go:117] "RemoveContainer" containerID="d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.502871 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klf5s\" (UniqueName: \"kubernetes.io/projected/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-kube-api-access-klf5s\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.502911 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.502925 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5426bc1-cb43-4c49-b2e5-b8ac43855c29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.507185 4957 scope.go:117] "RemoveContainer" containerID="517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f" Dec 06 06:36:28 crc kubenswrapper[4957]: E1206 06:36:28.507617 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f\": container with ID starting with 517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f not found: ID does not exist" containerID="517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.507650 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f"} err="failed to get container status \"517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f\": rpc error: code = NotFound desc = could not find container \"517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f\": container with ID starting with 517d7b55c29686b676c00d48aa1da103d1fd7185bd01612bd36ca936b775ae1f not found: ID does not exist" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.507670 4957 scope.go:117] "RemoveContainer" containerID="ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d" Dec 06 06:36:28 crc kubenswrapper[4957]: E1206 06:36:28.508024 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d\": container with ID starting with ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d not found: ID does not exist" containerID="ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.508110 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d"} err="failed to get container status \"ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d\": rpc error: code = NotFound desc = could not find container \"ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d\": container with ID starting with ee4d98dbce0fe5d5fa6af48fa0c7fc4742f188b11352dfa93f5de1ad5439347d not found: ID does not exist" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.508150 4957 scope.go:117] "RemoveContainer" containerID="d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b" Dec 06 06:36:28 crc kubenswrapper[4957]: E1206 06:36:28.508503 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b\": container with ID starting with d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b not found: ID does not exist" containerID="d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.508552 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b"} err="failed to get container status \"d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b\": rpc error: code = NotFound desc = could not find container \"d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b\": container with ID starting with d9d63fdff33f5eb312a9e398f39ac9550dc8cb54e4d2561cad0c358b84df728b not found: ID does not exist" Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.713042 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wd9"] Dec 06 06:36:28 crc kubenswrapper[4957]: I1206 06:36:28.724468 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f4wd9"] Dec 06 06:36:29 crc kubenswrapper[4957]: I1206 06:36:29.388577 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tts4c" event={"ID":"2903e894-9c88-41e4-abf4-9532e09cd605","Type":"ContainerStarted","Data":"a3a651d78c25d614207322d4c72c9965de52161a33eb5e5154055f9edfaae3d5"} Dec 06 06:36:29 crc kubenswrapper[4957]: I1206 06:36:29.408450 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tts4c" podStartSLOduration=3.007046397 podStartE2EDuration="5.408430121s" podCreationTimestamp="2025-12-06 06:36:24 +0000 UTC" firstStartedPulling="2025-12-06 06:36:26.351029673 +0000 UTC m=+3426.001297305" lastFinishedPulling="2025-12-06 06:36:28.752413397 +0000 UTC m=+3428.402681029" observedRunningTime="2025-12-06 06:36:29.40617069 +0000 UTC m=+3429.056438332" watchObservedRunningTime="2025-12-06 06:36:29.408430121 +0000 UTC m=+3429.058697763" Dec 06 06:36:30 crc kubenswrapper[4957]: I1206 06:36:30.672118 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" path="/var/lib/kubelet/pods/f5426bc1-cb43-4c49-b2e5-b8ac43855c29/volumes" Dec 06 06:36:34 crc kubenswrapper[4957]: I1206 06:36:34.656151 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:34 crc kubenswrapper[4957]: I1206 06:36:34.656677 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:34 crc kubenswrapper[4957]: I1206 06:36:34.728711 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:35 crc kubenswrapper[4957]: I1206 06:36:35.493529 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:35 crc kubenswrapper[4957]: I1206 06:36:35.538584 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tts4c"] Dec 06 06:36:37 crc kubenswrapper[4957]: I1206 06:36:37.460199 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tts4c" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="registry-server" containerID="cri-o://a3a651d78c25d614207322d4c72c9965de52161a33eb5e5154055f9edfaae3d5" gracePeriod=2 Dec 06 06:36:38 crc kubenswrapper[4957]: I1206 06:36:38.468490 4957 generic.go:334] "Generic (PLEG): container finished" podID="2903e894-9c88-41e4-abf4-9532e09cd605" containerID="a3a651d78c25d614207322d4c72c9965de52161a33eb5e5154055f9edfaae3d5" exitCode=0 Dec 06 06:36:38 crc kubenswrapper[4957]: I1206 06:36:38.468572 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tts4c" event={"ID":"2903e894-9c88-41e4-abf4-9532e09cd605","Type":"ContainerDied","Data":"a3a651d78c25d614207322d4c72c9965de52161a33eb5e5154055f9edfaae3d5"} Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.056647 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.131422 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-catalog-content\") pod \"2903e894-9c88-41e4-abf4-9532e09cd605\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.131517 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-utilities\") pod \"2903e894-9c88-41e4-abf4-9532e09cd605\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.131697 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84z8m\" (UniqueName: \"kubernetes.io/projected/2903e894-9c88-41e4-abf4-9532e09cd605-kube-api-access-84z8m\") pod \"2903e894-9c88-41e4-abf4-9532e09cd605\" (UID: \"2903e894-9c88-41e4-abf4-9532e09cd605\") " Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.132375 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-utilities" (OuterVolumeSpecName: "utilities") pod "2903e894-9c88-41e4-abf4-9532e09cd605" (UID: "2903e894-9c88-41e4-abf4-9532e09cd605"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.140344 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2903e894-9c88-41e4-abf4-9532e09cd605-kube-api-access-84z8m" (OuterVolumeSpecName: "kube-api-access-84z8m") pod "2903e894-9c88-41e4-abf4-9532e09cd605" (UID: "2903e894-9c88-41e4-abf4-9532e09cd605"). InnerVolumeSpecName "kube-api-access-84z8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.184898 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2903e894-9c88-41e4-abf4-9532e09cd605" (UID: "2903e894-9c88-41e4-abf4-9532e09cd605"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.234289 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.234328 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2903e894-9c88-41e4-abf4-9532e09cd605-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.234338 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84z8m\" (UniqueName: \"kubernetes.io/projected/2903e894-9c88-41e4-abf4-9532e09cd605-kube-api-access-84z8m\") on node \"crc\" DevicePath \"\"" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.479969 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tts4c" event={"ID":"2903e894-9c88-41e4-abf4-9532e09cd605","Type":"ContainerDied","Data":"e1b06a40c56fb52c84e76c343ce0d3118e416d380e72d551bca53e897082647f"} Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.480048 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tts4c" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.481134 4957 scope.go:117] "RemoveContainer" containerID="a3a651d78c25d614207322d4c72c9965de52161a33eb5e5154055f9edfaae3d5" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.499870 4957 scope.go:117] "RemoveContainer" containerID="bfd8d5749eb88035a3f0d9ef6708f0128cb886505611d27f1cce26e6820ba0ff" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.515323 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tts4c"] Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.522062 4957 scope.go:117] "RemoveContainer" containerID="af07479233ddd263e854288692111be7b407b29c41501df8f40697f1e25f7b88" Dec 06 06:36:39 crc kubenswrapper[4957]: I1206 06:36:39.524735 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tts4c"] Dec 06 06:36:40 crc kubenswrapper[4957]: I1206 06:36:40.181010 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:36:40 crc kubenswrapper[4957]: I1206 06:36:40.181486 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:36:40 crc kubenswrapper[4957]: I1206 06:36:40.677435 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" path="/var/lib/kubelet/pods/2903e894-9c88-41e4-abf4-9532e09cd605/volumes" Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.180924 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.181963 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.182055 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.183481 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4e3af660d7ea6175b7239e78bf6b4827a27406240b47d1ba1e1552fb3eca408"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.183568 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://d4e3af660d7ea6175b7239e78bf6b4827a27406240b47d1ba1e1552fb3eca408" gracePeriod=600 Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.773079 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="d4e3af660d7ea6175b7239e78bf6b4827a27406240b47d1ba1e1552fb3eca408" exitCode=0 Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.773170 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"d4e3af660d7ea6175b7239e78bf6b4827a27406240b47d1ba1e1552fb3eca408"} Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.774595 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508"} Dec 06 06:37:10 crc kubenswrapper[4957]: I1206 06:37:10.774624 4957 scope.go:117] "RemoveContainer" containerID="912ad788414e038af026f23f4f7485b30bc7c2d35499615fefb5249abbd4f536" Dec 06 06:39:10 crc kubenswrapper[4957]: I1206 06:39:10.181365 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:39:10 crc kubenswrapper[4957]: I1206 06:39:10.181955 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:39:40 crc kubenswrapper[4957]: I1206 06:39:40.181040 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:39:40 crc kubenswrapper[4957]: I1206 06:39:40.181728 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.181244 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.181879 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.181928 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.182699 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.182758 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" gracePeriod=600 Dec 06 06:40:10 crc kubenswrapper[4957]: E1206 06:40:10.306604 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.615210 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" exitCode=0 Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.615266 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508"} Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.615310 4957 scope.go:117] "RemoveContainer" containerID="d4e3af660d7ea6175b7239e78bf6b4827a27406240b47d1ba1e1552fb3eca408" Dec 06 06:40:10 crc kubenswrapper[4957]: I1206 06:40:10.616298 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:40:10 crc kubenswrapper[4957]: E1206 06:40:10.616996 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:40:22 crc kubenswrapper[4957]: I1206 06:40:22.662791 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:40:22 crc kubenswrapper[4957]: E1206 06:40:22.663763 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:40:35 crc kubenswrapper[4957]: I1206 06:40:35.663913 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:40:35 crc kubenswrapper[4957]: E1206 06:40:35.665349 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:40:46 crc kubenswrapper[4957]: I1206 06:40:46.663014 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:40:46 crc kubenswrapper[4957]: E1206 06:40:46.663945 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:40:57 crc kubenswrapper[4957]: I1206 06:40:57.663053 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:40:57 crc kubenswrapper[4957]: E1206 06:40:57.663784 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:41:09 crc kubenswrapper[4957]: I1206 06:41:09.663373 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:41:09 crc kubenswrapper[4957]: E1206 06:41:09.664182 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:41:24 crc kubenswrapper[4957]: I1206 06:41:24.662681 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:41:24 crc kubenswrapper[4957]: E1206 06:41:24.663445 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:41:37 crc kubenswrapper[4957]: I1206 06:41:37.661928 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:41:37 crc kubenswrapper[4957]: E1206 06:41:37.662595 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:41:48 crc kubenswrapper[4957]: I1206 06:41:48.662406 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:41:48 crc kubenswrapper[4957]: E1206 06:41:48.663293 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:42:02 crc kubenswrapper[4957]: I1206 06:42:02.662579 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:42:02 crc kubenswrapper[4957]: E1206 06:42:02.664281 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:42:17 crc kubenswrapper[4957]: I1206 06:42:17.663760 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:42:17 crc kubenswrapper[4957]: E1206 06:42:17.664877 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:42:30 crc kubenswrapper[4957]: I1206 06:42:30.672764 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:42:30 crc kubenswrapper[4957]: E1206 06:42:30.674336 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.172626 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mrb6w"] Dec 06 06:42:33 crc kubenswrapper[4957]: E1206 06:42:33.173419 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="registry-server" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173436 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="registry-server" Dec 06 06:42:33 crc kubenswrapper[4957]: E1206 06:42:33.173468 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="extract-content" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173476 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="extract-content" Dec 06 06:42:33 crc kubenswrapper[4957]: E1206 06:42:33.173489 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="extract-utilities" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173500 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="extract-utilities" Dec 06 06:42:33 crc kubenswrapper[4957]: E1206 06:42:33.173519 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="registry-server" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173527 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="registry-server" Dec 06 06:42:33 crc kubenswrapper[4957]: E1206 06:42:33.173546 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="extract-content" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173557 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="extract-content" Dec 06 06:42:33 crc kubenswrapper[4957]: E1206 06:42:33.173578 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="extract-utilities" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173589 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="extract-utilities" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173875 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5426bc1-cb43-4c49-b2e5-b8ac43855c29" containerName="registry-server" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.173894 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="2903e894-9c88-41e4-abf4-9532e09cd605" containerName="registry-server" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.175745 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.188948 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrb6w"] Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.271844 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-catalog-content\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.272017 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmz2d\" (UniqueName: \"kubernetes.io/projected/49d20cd6-2605-4a3e-9804-65f3d3566b4b-kube-api-access-cmz2d\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.272519 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-utilities\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.374694 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-utilities\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.374844 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-catalog-content\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.374889 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmz2d\" (UniqueName: \"kubernetes.io/projected/49d20cd6-2605-4a3e-9804-65f3d3566b4b-kube-api-access-cmz2d\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.375559 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-catalog-content\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.375970 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-utilities\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.395163 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmz2d\" (UniqueName: \"kubernetes.io/projected/49d20cd6-2605-4a3e-9804-65f3d3566b4b-kube-api-access-cmz2d\") pod \"certified-operators-mrb6w\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:33 crc kubenswrapper[4957]: I1206 06:42:33.502730 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:34 crc kubenswrapper[4957]: I1206 06:42:34.033637 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mrb6w"] Dec 06 06:42:34 crc kubenswrapper[4957]: I1206 06:42:34.949237 4957 generic.go:334] "Generic (PLEG): container finished" podID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerID="484503c5f9c2ee4b43d8b7e0785013d23ced9da2345a1a7ace5882cd0a0e6dba" exitCode=0 Dec 06 06:42:34 crc kubenswrapper[4957]: I1206 06:42:34.949392 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrb6w" event={"ID":"49d20cd6-2605-4a3e-9804-65f3d3566b4b","Type":"ContainerDied","Data":"484503c5f9c2ee4b43d8b7e0785013d23ced9da2345a1a7ace5882cd0a0e6dba"} Dec 06 06:42:34 crc kubenswrapper[4957]: I1206 06:42:34.949606 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrb6w" event={"ID":"49d20cd6-2605-4a3e-9804-65f3d3566b4b","Type":"ContainerStarted","Data":"b6028b267d56c5c364df045a1c76bc6be169dd4463ce661051a2882be8a9e978"} Dec 06 06:42:34 crc kubenswrapper[4957]: I1206 06:42:34.953216 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:42:36 crc kubenswrapper[4957]: I1206 06:42:36.968795 4957 generic.go:334] "Generic (PLEG): container finished" podID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerID="862d6842c81f4bf9510cf3ed6d63abdd8f33964f2a7aabdea1544b978e1745dd" exitCode=0 Dec 06 06:42:36 crc kubenswrapper[4957]: I1206 06:42:36.968909 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrb6w" event={"ID":"49d20cd6-2605-4a3e-9804-65f3d3566b4b","Type":"ContainerDied","Data":"862d6842c81f4bf9510cf3ed6d63abdd8f33964f2a7aabdea1544b978e1745dd"} Dec 06 06:42:38 crc kubenswrapper[4957]: I1206 06:42:38.988998 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrb6w" event={"ID":"49d20cd6-2605-4a3e-9804-65f3d3566b4b","Type":"ContainerStarted","Data":"d7ba271d8a053178a67d3618dc52be5a1b8d706bf2a8f2b670d8be5f0920a835"} Dec 06 06:42:39 crc kubenswrapper[4957]: I1206 06:42:39.008760 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mrb6w" podStartSLOduration=3.008389171 podStartE2EDuration="6.008741006s" podCreationTimestamp="2025-12-06 06:42:33 +0000 UTC" firstStartedPulling="2025-12-06 06:42:34.952786685 +0000 UTC m=+3794.603054327" lastFinishedPulling="2025-12-06 06:42:37.95313849 +0000 UTC m=+3797.603406162" observedRunningTime="2025-12-06 06:42:39.00410437 +0000 UTC m=+3798.654372022" watchObservedRunningTime="2025-12-06 06:42:39.008741006 +0000 UTC m=+3798.659008638" Dec 06 06:42:43 crc kubenswrapper[4957]: I1206 06:42:43.502950 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:43 crc kubenswrapper[4957]: I1206 06:42:43.503504 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:43 crc kubenswrapper[4957]: I1206 06:42:43.549283 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:43 crc kubenswrapper[4957]: I1206 06:42:43.666179 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:42:43 crc kubenswrapper[4957]: E1206 06:42:43.666713 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:42:44 crc kubenswrapper[4957]: I1206 06:42:44.069322 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:44 crc kubenswrapper[4957]: I1206 06:42:44.127292 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrb6w"] Dec 06 06:42:46 crc kubenswrapper[4957]: I1206 06:42:46.040667 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mrb6w" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="registry-server" containerID="cri-o://d7ba271d8a053178a67d3618dc52be5a1b8d706bf2a8f2b670d8be5f0920a835" gracePeriod=2 Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.052205 4957 generic.go:334] "Generic (PLEG): container finished" podID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerID="d7ba271d8a053178a67d3618dc52be5a1b8d706bf2a8f2b670d8be5f0920a835" exitCode=0 Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.052374 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrb6w" event={"ID":"49d20cd6-2605-4a3e-9804-65f3d3566b4b","Type":"ContainerDied","Data":"d7ba271d8a053178a67d3618dc52be5a1b8d706bf2a8f2b670d8be5f0920a835"} Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.193849 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.338860 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-utilities\") pod \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.338934 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmz2d\" (UniqueName: \"kubernetes.io/projected/49d20cd6-2605-4a3e-9804-65f3d3566b4b-kube-api-access-cmz2d\") pod \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.338984 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-catalog-content\") pod \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\" (UID: \"49d20cd6-2605-4a3e-9804-65f3d3566b4b\") " Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.340686 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-utilities" (OuterVolumeSpecName: "utilities") pod "49d20cd6-2605-4a3e-9804-65f3d3566b4b" (UID: "49d20cd6-2605-4a3e-9804-65f3d3566b4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.352073 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d20cd6-2605-4a3e-9804-65f3d3566b4b-kube-api-access-cmz2d" (OuterVolumeSpecName: "kube-api-access-cmz2d") pod "49d20cd6-2605-4a3e-9804-65f3d3566b4b" (UID: "49d20cd6-2605-4a3e-9804-65f3d3566b4b"). InnerVolumeSpecName "kube-api-access-cmz2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.441796 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.441851 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmz2d\" (UniqueName: \"kubernetes.io/projected/49d20cd6-2605-4a3e-9804-65f3d3566b4b-kube-api-access-cmz2d\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:47 crc kubenswrapper[4957]: I1206 06:42:47.998573 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49d20cd6-2605-4a3e-9804-65f3d3566b4b" (UID: "49d20cd6-2605-4a3e-9804-65f3d3566b4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.052080 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d20cd6-2605-4a3e-9804-65f3d3566b4b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.065431 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mrb6w" event={"ID":"49d20cd6-2605-4a3e-9804-65f3d3566b4b","Type":"ContainerDied","Data":"b6028b267d56c5c364df045a1c76bc6be169dd4463ce661051a2882be8a9e978"} Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.065489 4957 scope.go:117] "RemoveContainer" containerID="d7ba271d8a053178a67d3618dc52be5a1b8d706bf2a8f2b670d8be5f0920a835" Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.065514 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mrb6w" Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.095371 4957 scope.go:117] "RemoveContainer" containerID="862d6842c81f4bf9510cf3ed6d63abdd8f33964f2a7aabdea1544b978e1745dd" Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.101390 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mrb6w"] Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.112021 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mrb6w"] Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.139058 4957 scope.go:117] "RemoveContainer" containerID="484503c5f9c2ee4b43d8b7e0785013d23ced9da2345a1a7ace5882cd0a0e6dba" Dec 06 06:42:48 crc kubenswrapper[4957]: I1206 06:42:48.677228 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" path="/var/lib/kubelet/pods/49d20cd6-2605-4a3e-9804-65f3d3566b4b/volumes" Dec 06 06:42:58 crc kubenswrapper[4957]: I1206 06:42:58.662789 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:42:58 crc kubenswrapper[4957]: E1206 06:42:58.664536 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:43:11 crc kubenswrapper[4957]: I1206 06:43:11.662600 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:43:11 crc kubenswrapper[4957]: E1206 06:43:11.664791 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:43:21 crc kubenswrapper[4957]: I1206 06:43:21.382681 4957 generic.go:334] "Generic (PLEG): container finished" podID="469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" containerID="2d93e442b77f6a1c41b471530fa69a446fe5b5b61d3c33320d7876e25efc2f5c" exitCode=0 Dec 06 06:43:21 crc kubenswrapper[4957]: I1206 06:43:21.382803 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd","Type":"ContainerDied","Data":"2d93e442b77f6a1c41b471530fa69a446fe5b5b61d3c33320d7876e25efc2f5c"} Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.772988 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.865781 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.865909 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-temporary\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.865997 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-config-data\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.866111 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd7qq\" (UniqueName: \"kubernetes.io/projected/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-kube-api-access-wd7qq\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.866138 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config-secret\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.866196 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.866318 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ca-certs\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.866395 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ssh-key\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.866481 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-workdir\") pod \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\" (UID: \"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd\") " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.867262 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.867434 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-config-data" (OuterVolumeSpecName: "config-data") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.867774 4957 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.867803 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.872023 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.873801 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.889067 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-kube-api-access-wd7qq" (OuterVolumeSpecName: "kube-api-access-wd7qq") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "kube-api-access-wd7qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.925431 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.925567 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.930256 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.951915 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" (UID: "469a494f-1ceb-4d9f-bfe5-e0856e88f5bd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.969547 4957 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.969581 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd7qq\" (UniqueName: \"kubernetes.io/projected/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-kube-api-access-wd7qq\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.969593 4957 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.969625 4957 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.969635 4957 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.969643 4957 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.969655 4957 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/469a494f-1ceb-4d9f-bfe5-e0856e88f5bd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:22 crc kubenswrapper[4957]: I1206 06:43:22.989026 4957 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 06 06:43:23 crc kubenswrapper[4957]: I1206 06:43:23.071616 4957 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 06 06:43:23 crc kubenswrapper[4957]: I1206 06:43:23.400954 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"469a494f-1ceb-4d9f-bfe5-e0856e88f5bd","Type":"ContainerDied","Data":"abc796ec6778cae473e15245a2df316c5f6481db47a112062264f412314de8da"} Dec 06 06:43:23 crc kubenswrapper[4957]: I1206 06:43:23.401266 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abc796ec6778cae473e15245a2df316c5f6481db47a112062264f412314de8da" Dec 06 06:43:23 crc kubenswrapper[4957]: I1206 06:43:23.401342 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 06:43:25 crc kubenswrapper[4957]: I1206 06:43:25.662898 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:43:25 crc kubenswrapper[4957]: E1206 06:43:25.664042 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.783657 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 06:43:34 crc kubenswrapper[4957]: E1206 06:43:34.784630 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="registry-server" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.784644 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="registry-server" Dec 06 06:43:34 crc kubenswrapper[4957]: E1206 06:43:34.784700 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" containerName="tempest-tests-tempest-tests-runner" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.784712 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" containerName="tempest-tests-tempest-tests-runner" Dec 06 06:43:34 crc kubenswrapper[4957]: E1206 06:43:34.784723 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="extract-content" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.784729 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="extract-content" Dec 06 06:43:34 crc kubenswrapper[4957]: E1206 06:43:34.784743 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="extract-utilities" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.784749 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="extract-utilities" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.785002 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="49d20cd6-2605-4a3e-9804-65f3d3566b4b" containerName="registry-server" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.785033 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="469a494f-1ceb-4d9f-bfe5-e0856e88f5bd" containerName="tempest-tests-tempest-tests-runner" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.785798 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.789606 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-m7bqq" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.793944 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.835820 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbhvc\" (UniqueName: \"kubernetes.io/projected/d42a9079-5c1e-424a-b79b-d469afb272c3-kube-api-access-zbhvc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d42a9079-5c1e-424a-b79b-d469afb272c3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.835915 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d42a9079-5c1e-424a-b79b-d469afb272c3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.937176 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbhvc\" (UniqueName: \"kubernetes.io/projected/d42a9079-5c1e-424a-b79b-d469afb272c3-kube-api-access-zbhvc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d42a9079-5c1e-424a-b79b-d469afb272c3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.937237 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d42a9079-5c1e-424a-b79b-d469afb272c3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.937733 4957 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d42a9079-5c1e-424a-b79b-d469afb272c3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.964327 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbhvc\" (UniqueName: \"kubernetes.io/projected/d42a9079-5c1e-424a-b79b-d469afb272c3-kube-api-access-zbhvc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d42a9079-5c1e-424a-b79b-d469afb272c3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:34 crc kubenswrapper[4957]: I1206 06:43:34.965296 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d42a9079-5c1e-424a-b79b-d469afb272c3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:35 crc kubenswrapper[4957]: I1206 06:43:35.109699 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 06:43:35 crc kubenswrapper[4957]: I1206 06:43:35.554881 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 06:43:36 crc kubenswrapper[4957]: I1206 06:43:36.512392 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d42a9079-5c1e-424a-b79b-d469afb272c3","Type":"ContainerStarted","Data":"608349055d202be8574460c3ddfac188245cd07036244dbd22ce32599abdb2ee"} Dec 06 06:43:37 crc kubenswrapper[4957]: I1206 06:43:37.524753 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d42a9079-5c1e-424a-b79b-d469afb272c3","Type":"ContainerStarted","Data":"379d6677ec501c9bd66cef7946ba5843eb2b1cf34908230d61808cd0eafe143c"} Dec 06 06:43:37 crc kubenswrapper[4957]: I1206 06:43:37.541347 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.502415899 podStartE2EDuration="3.541329013s" podCreationTimestamp="2025-12-06 06:43:34 +0000 UTC" firstStartedPulling="2025-12-06 06:43:35.561821631 +0000 UTC m=+3855.212089283" lastFinishedPulling="2025-12-06 06:43:36.600734765 +0000 UTC m=+3856.251002397" observedRunningTime="2025-12-06 06:43:37.538014813 +0000 UTC m=+3857.188282485" watchObservedRunningTime="2025-12-06 06:43:37.541329013 +0000 UTC m=+3857.191596655" Dec 06 06:43:40 crc kubenswrapper[4957]: I1206 06:43:40.670811 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:43:40 crc kubenswrapper[4957]: E1206 06:43:40.672013 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:43:55 crc kubenswrapper[4957]: I1206 06:43:55.662927 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:43:55 crc kubenswrapper[4957]: E1206 06:43:55.663820 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.102813 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vlspc/must-gather-sbsvd"] Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.105064 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.106924 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vlspc"/"kube-root-ca.crt" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.106923 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vlspc"/"default-dockercfg-4wkld" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.106982 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vlspc"/"openshift-service-ca.crt" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.110855 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vlspc/must-gather-sbsvd"] Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.213646 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a3a3017f-6621-4931-80cd-040dc7a81d9f-must-gather-output\") pod \"must-gather-sbsvd\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.214071 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bpw\" (UniqueName: \"kubernetes.io/projected/a3a3017f-6621-4931-80cd-040dc7a81d9f-kube-api-access-v6bpw\") pod \"must-gather-sbsvd\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.315687 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a3a3017f-6621-4931-80cd-040dc7a81d9f-must-gather-output\") pod \"must-gather-sbsvd\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.315800 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6bpw\" (UniqueName: \"kubernetes.io/projected/a3a3017f-6621-4931-80cd-040dc7a81d9f-kube-api-access-v6bpw\") pod \"must-gather-sbsvd\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.316225 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a3a3017f-6621-4931-80cd-040dc7a81d9f-must-gather-output\") pod \"must-gather-sbsvd\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.336760 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6bpw\" (UniqueName: \"kubernetes.io/projected/a3a3017f-6621-4931-80cd-040dc7a81d9f-kube-api-access-v6bpw\") pod \"must-gather-sbsvd\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.428869 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:44:02 crc kubenswrapper[4957]: I1206 06:44:02.875921 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vlspc/must-gather-sbsvd"] Dec 06 06:44:03 crc kubenswrapper[4957]: I1206 06:44:03.783584 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/must-gather-sbsvd" event={"ID":"a3a3017f-6621-4931-80cd-040dc7a81d9f","Type":"ContainerStarted","Data":"306dcbbec36e616d92c625560b08a3983bc36dc55b374d7e547fcc08c8a8da1b"} Dec 06 06:44:06 crc kubenswrapper[4957]: I1206 06:44:06.663288 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:44:06 crc kubenswrapper[4957]: E1206 06:44:06.664563 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:44:07 crc kubenswrapper[4957]: I1206 06:44:07.820597 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/must-gather-sbsvd" event={"ID":"a3a3017f-6621-4931-80cd-040dc7a81d9f","Type":"ContainerStarted","Data":"76ded26ea0e3694c4ad59de8bd22b9028413b3e29389b7eee8e441ab3272a63f"} Dec 06 06:44:07 crc kubenswrapper[4957]: I1206 06:44:07.820969 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/must-gather-sbsvd" event={"ID":"a3a3017f-6621-4931-80cd-040dc7a81d9f","Type":"ContainerStarted","Data":"c29bfa67dfacfb218cff8778e68c95cb16e9bd8b46df8e07034fbcbd4865c741"} Dec 06 06:44:07 crc kubenswrapper[4957]: I1206 06:44:07.840909 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vlspc/must-gather-sbsvd" podStartSLOduration=1.873065159 podStartE2EDuration="5.8408876s" podCreationTimestamp="2025-12-06 06:44:02 +0000 UTC" firstStartedPulling="2025-12-06 06:44:02.885037655 +0000 UTC m=+3882.535305287" lastFinishedPulling="2025-12-06 06:44:06.852860106 +0000 UTC m=+3886.503127728" observedRunningTime="2025-12-06 06:44:07.833689415 +0000 UTC m=+3887.483957057" watchObservedRunningTime="2025-12-06 06:44:07.8408876 +0000 UTC m=+3887.491155232" Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.676046 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vlspc/crc-debug-7xjpv"] Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.678385 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.777203 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-host\") pod \"crc-debug-7xjpv\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.777283 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnnr2\" (UniqueName: \"kubernetes.io/projected/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-kube-api-access-rnnr2\") pod \"crc-debug-7xjpv\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.878646 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-host\") pod \"crc-debug-7xjpv\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.878709 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnnr2\" (UniqueName: \"kubernetes.io/projected/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-kube-api-access-rnnr2\") pod \"crc-debug-7xjpv\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.878780 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-host\") pod \"crc-debug-7xjpv\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:10 crc kubenswrapper[4957]: I1206 06:44:10.899590 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnnr2\" (UniqueName: \"kubernetes.io/projected/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-kube-api-access-rnnr2\") pod \"crc-debug-7xjpv\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:11 crc kubenswrapper[4957]: I1206 06:44:11.000551 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:44:11 crc kubenswrapper[4957]: W1206 06:44:11.048114 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc5c2fb1_9ffa_40ae_b517_3e7795ff3632.slice/crio-612575d472370fb4dd6717ca9e6b9980420aba992214451ed568e8adce250312 WatchSource:0}: Error finding container 612575d472370fb4dd6717ca9e6b9980420aba992214451ed568e8adce250312: Status 404 returned error can't find the container with id 612575d472370fb4dd6717ca9e6b9980420aba992214451ed568e8adce250312 Dec 06 06:44:11 crc kubenswrapper[4957]: I1206 06:44:11.858701 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" event={"ID":"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632","Type":"ContainerStarted","Data":"612575d472370fb4dd6717ca9e6b9980420aba992214451ed568e8adce250312"} Dec 06 06:44:21 crc kubenswrapper[4957]: I1206 06:44:21.663004 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:44:22 crc kubenswrapper[4957]: E1206 06:44:21.664141 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:44:23 crc kubenswrapper[4957]: I1206 06:44:23.992064 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" event={"ID":"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632","Type":"ContainerStarted","Data":"4a29b460acf8d10d348da4dba51b8ef85495afaa2f1638aaf3157d65dbb196f9"} Dec 06 06:44:24 crc kubenswrapper[4957]: I1206 06:44:24.005994 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" podStartSLOduration=1.8663424549999998 podStartE2EDuration="14.005977583s" podCreationTimestamp="2025-12-06 06:44:10 +0000 UTC" firstStartedPulling="2025-12-06 06:44:11.050331186 +0000 UTC m=+3890.700598818" lastFinishedPulling="2025-12-06 06:44:23.189966314 +0000 UTC m=+3902.840233946" observedRunningTime="2025-12-06 06:44:24.00400374 +0000 UTC m=+3903.654271392" watchObservedRunningTime="2025-12-06 06:44:24.005977583 +0000 UTC m=+3903.656245215" Dec 06 06:44:33 crc kubenswrapper[4957]: I1206 06:44:33.661824 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:44:33 crc kubenswrapper[4957]: E1206 06:44:33.662729 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:44:44 crc kubenswrapper[4957]: I1206 06:44:44.663100 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:44:44 crc kubenswrapper[4957]: E1206 06:44:44.663935 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:44:56 crc kubenswrapper[4957]: I1206 06:44:56.662950 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:44:56 crc kubenswrapper[4957]: E1206 06:44:56.663760 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.178226 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5"] Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.179942 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.183663 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.189987 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.190256 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5"] Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.191197 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msjwf\" (UniqueName: \"kubernetes.io/projected/469d309d-5b5d-4c02-b324-b07d63ff5942-kube-api-access-msjwf\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.191446 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/469d309d-5b5d-4c02-b324-b07d63ff5942-config-volume\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.191661 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/469d309d-5b5d-4c02-b324-b07d63ff5942-secret-volume\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.293236 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msjwf\" (UniqueName: \"kubernetes.io/projected/469d309d-5b5d-4c02-b324-b07d63ff5942-kube-api-access-msjwf\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.293369 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/469d309d-5b5d-4c02-b324-b07d63ff5942-config-volume\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.293505 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/469d309d-5b5d-4c02-b324-b07d63ff5942-secret-volume\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.294345 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/469d309d-5b5d-4c02-b324-b07d63ff5942-config-volume\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.483786 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msjwf\" (UniqueName: \"kubernetes.io/projected/469d309d-5b5d-4c02-b324-b07d63ff5942-kube-api-access-msjwf\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.483795 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/469d309d-5b5d-4c02-b324-b07d63ff5942-secret-volume\") pod \"collect-profiles-29416725-vmpt5\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:00 crc kubenswrapper[4957]: I1206 06:45:00.501792 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:01 crc kubenswrapper[4957]: I1206 06:45:01.008238 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5"] Dec 06 06:45:01 crc kubenswrapper[4957]: I1206 06:45:01.309798 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" event={"ID":"469d309d-5b5d-4c02-b324-b07d63ff5942","Type":"ContainerStarted","Data":"5fca80cbd745c1a61b0b6f83f22ad55748003e8b8b7ed1d3050fdad5b3d55aa1"} Dec 06 06:45:01 crc kubenswrapper[4957]: I1206 06:45:01.310186 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" event={"ID":"469d309d-5b5d-4c02-b324-b07d63ff5942","Type":"ContainerStarted","Data":"30d999bec6ecae1ab9355ebd7c99e3e175c133f55822da3e8a50e05d0e137e86"} Dec 06 06:45:01 crc kubenswrapper[4957]: I1206 06:45:01.334921 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" podStartSLOduration=1.334896209 podStartE2EDuration="1.334896209s" podCreationTimestamp="2025-12-06 06:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:45:01.327789816 +0000 UTC m=+3940.978057468" watchObservedRunningTime="2025-12-06 06:45:01.334896209 +0000 UTC m=+3940.985163851" Dec 06 06:45:02 crc kubenswrapper[4957]: I1206 06:45:02.319444 4957 generic.go:334] "Generic (PLEG): container finished" podID="469d309d-5b5d-4c02-b324-b07d63ff5942" containerID="5fca80cbd745c1a61b0b6f83f22ad55748003e8b8b7ed1d3050fdad5b3d55aa1" exitCode=0 Dec 06 06:45:02 crc kubenswrapper[4957]: I1206 06:45:02.319621 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" event={"ID":"469d309d-5b5d-4c02-b324-b07d63ff5942","Type":"ContainerDied","Data":"5fca80cbd745c1a61b0b6f83f22ad55748003e8b8b7ed1d3050fdad5b3d55aa1"} Dec 06 06:45:03 crc kubenswrapper[4957]: I1206 06:45:03.758197 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:03 crc kubenswrapper[4957]: I1206 06:45:03.901793 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msjwf\" (UniqueName: \"kubernetes.io/projected/469d309d-5b5d-4c02-b324-b07d63ff5942-kube-api-access-msjwf\") pod \"469d309d-5b5d-4c02-b324-b07d63ff5942\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " Dec 06 06:45:03 crc kubenswrapper[4957]: I1206 06:45:03.902374 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/469d309d-5b5d-4c02-b324-b07d63ff5942-secret-volume\") pod \"469d309d-5b5d-4c02-b324-b07d63ff5942\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " Dec 06 06:45:03 crc kubenswrapper[4957]: I1206 06:45:03.902492 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/469d309d-5b5d-4c02-b324-b07d63ff5942-config-volume\") pod \"469d309d-5b5d-4c02-b324-b07d63ff5942\" (UID: \"469d309d-5b5d-4c02-b324-b07d63ff5942\") " Dec 06 06:45:03 crc kubenswrapper[4957]: I1206 06:45:03.903665 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/469d309d-5b5d-4c02-b324-b07d63ff5942-config-volume" (OuterVolumeSpecName: "config-volume") pod "469d309d-5b5d-4c02-b324-b07d63ff5942" (UID: "469d309d-5b5d-4c02-b324-b07d63ff5942"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 06:45:03 crc kubenswrapper[4957]: I1206 06:45:03.908094 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469d309d-5b5d-4c02-b324-b07d63ff5942-kube-api-access-msjwf" (OuterVolumeSpecName: "kube-api-access-msjwf") pod "469d309d-5b5d-4c02-b324-b07d63ff5942" (UID: "469d309d-5b5d-4c02-b324-b07d63ff5942"). InnerVolumeSpecName "kube-api-access-msjwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:45:03 crc kubenswrapper[4957]: I1206 06:45:03.916057 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469d309d-5b5d-4c02-b324-b07d63ff5942-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "469d309d-5b5d-4c02-b324-b07d63ff5942" (UID: "469d309d-5b5d-4c02-b324-b07d63ff5942"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.004526 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msjwf\" (UniqueName: \"kubernetes.io/projected/469d309d-5b5d-4c02-b324-b07d63ff5942-kube-api-access-msjwf\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.004563 4957 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/469d309d-5b5d-4c02-b324-b07d63ff5942-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.004576 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/469d309d-5b5d-4c02-b324-b07d63ff5942-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.340470 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" event={"ID":"469d309d-5b5d-4c02-b324-b07d63ff5942","Type":"ContainerDied","Data":"30d999bec6ecae1ab9355ebd7c99e3e175c133f55822da3e8a50e05d0e137e86"} Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.340510 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30d999bec6ecae1ab9355ebd7c99e3e175c133f55822da3e8a50e05d0e137e86" Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.340565 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416725-vmpt5" Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.411663 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk"] Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.422904 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416680-tn5dk"] Dec 06 06:45:04 crc kubenswrapper[4957]: I1206 06:45:04.673128 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4998358d-ec92-4690-b968-23c82a67a764" path="/var/lib/kubelet/pods/4998358d-ec92-4690-b968-23c82a67a764/volumes" Dec 06 06:45:06 crc kubenswrapper[4957]: I1206 06:45:06.361558 4957 generic.go:334] "Generic (PLEG): container finished" podID="dc5c2fb1-9ffa-40ae-b517-3e7795ff3632" containerID="4a29b460acf8d10d348da4dba51b8ef85495afaa2f1638aaf3157d65dbb196f9" exitCode=0 Dec 06 06:45:06 crc kubenswrapper[4957]: I1206 06:45:06.361666 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" event={"ID":"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632","Type":"ContainerDied","Data":"4a29b460acf8d10d348da4dba51b8ef85495afaa2f1638aaf3157d65dbb196f9"} Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.515785 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.547327 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vlspc/crc-debug-7xjpv"] Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.554349 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vlspc/crc-debug-7xjpv"] Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.662618 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:45:07 crc kubenswrapper[4957]: E1206 06:45:07.662903 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.680970 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-host\") pod \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.681325 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnnr2\" (UniqueName: \"kubernetes.io/projected/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-kube-api-access-rnnr2\") pod \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\" (UID: \"dc5c2fb1-9ffa-40ae-b517-3e7795ff3632\") " Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.681195 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-host" (OuterVolumeSpecName: "host") pod "dc5c2fb1-9ffa-40ae-b517-3e7795ff3632" (UID: "dc5c2fb1-9ffa-40ae-b517-3e7795ff3632"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:45:07 crc kubenswrapper[4957]: I1206 06:45:07.682306 4957 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-host\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.177206 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-kube-api-access-rnnr2" (OuterVolumeSpecName: "kube-api-access-rnnr2") pod "dc5c2fb1-9ffa-40ae-b517-3e7795ff3632" (UID: "dc5c2fb1-9ffa-40ae-b517-3e7795ff3632"). InnerVolumeSpecName "kube-api-access-rnnr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.191074 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnnr2\" (UniqueName: \"kubernetes.io/projected/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632-kube-api-access-rnnr2\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.380742 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="612575d472370fb4dd6717ca9e6b9980420aba992214451ed568e8adce250312" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.380825 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-7xjpv" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.673876 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc5c2fb1-9ffa-40ae-b517-3e7795ff3632" path="/var/lib/kubelet/pods/dc5c2fb1-9ffa-40ae-b517-3e7795ff3632/volumes" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.748148 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vlspc/crc-debug-gkx4p"] Dec 06 06:45:08 crc kubenswrapper[4957]: E1206 06:45:08.749033 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5c2fb1-9ffa-40ae-b517-3e7795ff3632" containerName="container-00" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.749060 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5c2fb1-9ffa-40ae-b517-3e7795ff3632" containerName="container-00" Dec 06 06:45:08 crc kubenswrapper[4957]: E1206 06:45:08.749074 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469d309d-5b5d-4c02-b324-b07d63ff5942" containerName="collect-profiles" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.749083 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="469d309d-5b5d-4c02-b324-b07d63ff5942" containerName="collect-profiles" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.749297 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5c2fb1-9ffa-40ae-b517-3e7795ff3632" containerName="container-00" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.749340 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="469d309d-5b5d-4c02-b324-b07d63ff5942" containerName="collect-profiles" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.750263 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.906777 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7fqk\" (UniqueName: \"kubernetes.io/projected/6bba5ebb-441a-4171-a865-6c8692b5f439-kube-api-access-p7fqk\") pod \"crc-debug-gkx4p\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:08 crc kubenswrapper[4957]: I1206 06:45:08.907482 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6bba5ebb-441a-4171-a865-6c8692b5f439-host\") pod \"crc-debug-gkx4p\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:09 crc kubenswrapper[4957]: I1206 06:45:09.010530 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7fqk\" (UniqueName: \"kubernetes.io/projected/6bba5ebb-441a-4171-a865-6c8692b5f439-kube-api-access-p7fqk\") pod \"crc-debug-gkx4p\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:09 crc kubenswrapper[4957]: I1206 06:45:09.010774 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6bba5ebb-441a-4171-a865-6c8692b5f439-host\") pod \"crc-debug-gkx4p\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:09 crc kubenswrapper[4957]: I1206 06:45:09.011079 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6bba5ebb-441a-4171-a865-6c8692b5f439-host\") pod \"crc-debug-gkx4p\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:09 crc kubenswrapper[4957]: I1206 06:45:09.034082 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7fqk\" (UniqueName: \"kubernetes.io/projected/6bba5ebb-441a-4171-a865-6c8692b5f439-kube-api-access-p7fqk\") pod \"crc-debug-gkx4p\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:09 crc kubenswrapper[4957]: I1206 06:45:09.067780 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:09 crc kubenswrapper[4957]: I1206 06:45:09.391076 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" event={"ID":"6bba5ebb-441a-4171-a865-6c8692b5f439","Type":"ContainerStarted","Data":"ddfde781d2f91e129833b5fc01bbd03377682f2bd0742538383738aa28fa7db3"} Dec 06 06:45:10 crc kubenswrapper[4957]: I1206 06:45:10.403955 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" event={"ID":"6bba5ebb-441a-4171-a865-6c8692b5f439","Type":"ContainerStarted","Data":"e9598d2687176b62df997a1fd3c35a95d04fe4bd09ebc1496bf5e9a6ea664faa"} Dec 06 06:45:10 crc kubenswrapper[4957]: I1206 06:45:10.427635 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" podStartSLOduration=2.427609318 podStartE2EDuration="2.427609318s" podCreationTimestamp="2025-12-06 06:45:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:45:10.417991728 +0000 UTC m=+3950.068259380" watchObservedRunningTime="2025-12-06 06:45:10.427609318 +0000 UTC m=+3950.077876940" Dec 06 06:45:11 crc kubenswrapper[4957]: I1206 06:45:11.415009 4957 generic.go:334] "Generic (PLEG): container finished" podID="6bba5ebb-441a-4171-a865-6c8692b5f439" containerID="e9598d2687176b62df997a1fd3c35a95d04fe4bd09ebc1496bf5e9a6ea664faa" exitCode=0 Dec 06 06:45:11 crc kubenswrapper[4957]: I1206 06:45:11.415079 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" event={"ID":"6bba5ebb-441a-4171-a865-6c8692b5f439","Type":"ContainerDied","Data":"e9598d2687176b62df997a1fd3c35a95d04fe4bd09ebc1496bf5e9a6ea664faa"} Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.544454 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.580522 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vlspc/crc-debug-gkx4p"] Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.591293 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vlspc/crc-debug-gkx4p"] Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.682034 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7fqk\" (UniqueName: \"kubernetes.io/projected/6bba5ebb-441a-4171-a865-6c8692b5f439-kube-api-access-p7fqk\") pod \"6bba5ebb-441a-4171-a865-6c8692b5f439\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.682094 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6bba5ebb-441a-4171-a865-6c8692b5f439-host\") pod \"6bba5ebb-441a-4171-a865-6c8692b5f439\" (UID: \"6bba5ebb-441a-4171-a865-6c8692b5f439\") " Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.682297 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bba5ebb-441a-4171-a865-6c8692b5f439-host" (OuterVolumeSpecName: "host") pod "6bba5ebb-441a-4171-a865-6c8692b5f439" (UID: "6bba5ebb-441a-4171-a865-6c8692b5f439"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.682618 4957 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6bba5ebb-441a-4171-a865-6c8692b5f439-host\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.688088 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bba5ebb-441a-4171-a865-6c8692b5f439-kube-api-access-p7fqk" (OuterVolumeSpecName: "kube-api-access-p7fqk") pod "6bba5ebb-441a-4171-a865-6c8692b5f439" (UID: "6bba5ebb-441a-4171-a865-6c8692b5f439"). InnerVolumeSpecName "kube-api-access-p7fqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:45:12 crc kubenswrapper[4957]: I1206 06:45:12.784164 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7fqk\" (UniqueName: \"kubernetes.io/projected/6bba5ebb-441a-4171-a865-6c8692b5f439-kube-api-access-p7fqk\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.433799 4957 scope.go:117] "RemoveContainer" containerID="e9598d2687176b62df997a1fd3c35a95d04fe4bd09ebc1496bf5e9a6ea664faa" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.433922 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-gkx4p" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.770053 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vlspc/crc-debug-lzsjw"] Dec 06 06:45:13 crc kubenswrapper[4957]: E1206 06:45:13.770749 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bba5ebb-441a-4171-a865-6c8692b5f439" containerName="container-00" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.770768 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bba5ebb-441a-4171-a865-6c8692b5f439" containerName="container-00" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.771048 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bba5ebb-441a-4171-a865-6c8692b5f439" containerName="container-00" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.771782 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.804230 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxqs8\" (UniqueName: \"kubernetes.io/projected/c71d7471-1826-4dab-8b23-ce3ab05a5e75-kube-api-access-fxqs8\") pod \"crc-debug-lzsjw\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.804308 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c71d7471-1826-4dab-8b23-ce3ab05a5e75-host\") pod \"crc-debug-lzsjw\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.906505 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxqs8\" (UniqueName: \"kubernetes.io/projected/c71d7471-1826-4dab-8b23-ce3ab05a5e75-kube-api-access-fxqs8\") pod \"crc-debug-lzsjw\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.906582 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c71d7471-1826-4dab-8b23-ce3ab05a5e75-host\") pod \"crc-debug-lzsjw\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.906679 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c71d7471-1826-4dab-8b23-ce3ab05a5e75-host\") pod \"crc-debug-lzsjw\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:13 crc kubenswrapper[4957]: I1206 06:45:13.926281 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxqs8\" (UniqueName: \"kubernetes.io/projected/c71d7471-1826-4dab-8b23-ce3ab05a5e75-kube-api-access-fxqs8\") pod \"crc-debug-lzsjw\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:14 crc kubenswrapper[4957]: I1206 06:45:14.099167 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:14 crc kubenswrapper[4957]: I1206 06:45:14.443856 4957 generic.go:334] "Generic (PLEG): container finished" podID="c71d7471-1826-4dab-8b23-ce3ab05a5e75" containerID="6a8c1d81833c4fd563551479cbacd8f15163705d863ebaeec49b0daa525b11c3" exitCode=0 Dec 06 06:45:14 crc kubenswrapper[4957]: I1206 06:45:14.443955 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-lzsjw" event={"ID":"c71d7471-1826-4dab-8b23-ce3ab05a5e75","Type":"ContainerDied","Data":"6a8c1d81833c4fd563551479cbacd8f15163705d863ebaeec49b0daa525b11c3"} Dec 06 06:45:14 crc kubenswrapper[4957]: I1206 06:45:14.444248 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/crc-debug-lzsjw" event={"ID":"c71d7471-1826-4dab-8b23-ce3ab05a5e75","Type":"ContainerStarted","Data":"55df35c41cf707247adc8e8a46dcfe1c2371837fa841a4eadec91288689b4f63"} Dec 06 06:45:14 crc kubenswrapper[4957]: I1206 06:45:14.484858 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vlspc/crc-debug-lzsjw"] Dec 06 06:45:14 crc kubenswrapper[4957]: I1206 06:45:14.494009 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vlspc/crc-debug-lzsjw"] Dec 06 06:45:14 crc kubenswrapper[4957]: I1206 06:45:14.675342 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bba5ebb-441a-4171-a865-6c8692b5f439" path="/var/lib/kubelet/pods/6bba5ebb-441a-4171-a865-6c8692b5f439/volumes" Dec 06 06:45:15 crc kubenswrapper[4957]: I1206 06:45:15.565711 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:15 crc kubenswrapper[4957]: I1206 06:45:15.643011 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c71d7471-1826-4dab-8b23-ce3ab05a5e75-host\") pod \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " Dec 06 06:45:15 crc kubenswrapper[4957]: I1206 06:45:15.643138 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c71d7471-1826-4dab-8b23-ce3ab05a5e75-host" (OuterVolumeSpecName: "host") pod "c71d7471-1826-4dab-8b23-ce3ab05a5e75" (UID: "c71d7471-1826-4dab-8b23-ce3ab05a5e75"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:45:15 crc kubenswrapper[4957]: I1206 06:45:15.643171 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxqs8\" (UniqueName: \"kubernetes.io/projected/c71d7471-1826-4dab-8b23-ce3ab05a5e75-kube-api-access-fxqs8\") pod \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\" (UID: \"c71d7471-1826-4dab-8b23-ce3ab05a5e75\") " Dec 06 06:45:15 crc kubenswrapper[4957]: I1206 06:45:15.643911 4957 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c71d7471-1826-4dab-8b23-ce3ab05a5e75-host\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:15 crc kubenswrapper[4957]: I1206 06:45:15.650068 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c71d7471-1826-4dab-8b23-ce3ab05a5e75-kube-api-access-fxqs8" (OuterVolumeSpecName: "kube-api-access-fxqs8") pod "c71d7471-1826-4dab-8b23-ce3ab05a5e75" (UID: "c71d7471-1826-4dab-8b23-ce3ab05a5e75"). InnerVolumeSpecName "kube-api-access-fxqs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:45:15 crc kubenswrapper[4957]: I1206 06:45:15.747057 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxqs8\" (UniqueName: \"kubernetes.io/projected/c71d7471-1826-4dab-8b23-ce3ab05a5e75-kube-api-access-fxqs8\") on node \"crc\" DevicePath \"\"" Dec 06 06:45:16 crc kubenswrapper[4957]: I1206 06:45:16.480182 4957 scope.go:117] "RemoveContainer" containerID="6a8c1d81833c4fd563551479cbacd8f15163705d863ebaeec49b0daa525b11c3" Dec 06 06:45:16 crc kubenswrapper[4957]: I1206 06:45:16.480324 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/crc-debug-lzsjw" Dec 06 06:45:16 crc kubenswrapper[4957]: I1206 06:45:16.672881 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c71d7471-1826-4dab-8b23-ce3ab05a5e75" path="/var/lib/kubelet/pods/c71d7471-1826-4dab-8b23-ce3ab05a5e75/volumes" Dec 06 06:45:22 crc kubenswrapper[4957]: I1206 06:45:22.661978 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:45:23 crc kubenswrapper[4957]: I1206 06:45:23.544092 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"6ccddb8ad5bca662c4e56f40c21bfbea3eb92dbae28847228b123252634f9d79"} Dec 06 06:45:31 crc kubenswrapper[4957]: I1206 06:45:31.665544 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c9f9fb958-p7bk5_59b9c817-9c5c-4f08-9ca9-9d54e1d0912b/barbican-api/0.log" Dec 06 06:45:31 crc kubenswrapper[4957]: I1206 06:45:31.831217 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c9f9fb958-p7bk5_59b9c817-9c5c-4f08-9ca9-9d54e1d0912b/barbican-api-log/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.005366 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bb9fc9b94-kknvp_e3907b98-0bd2-4ad9-bb72-d1f651c16f68/barbican-keystone-listener/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.071580 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bb9fc9b94-kknvp_e3907b98-0bd2-4ad9-bb72-d1f651c16f68/barbican-keystone-listener-log/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.197460 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5684df76ff-6tw89_b1c9d4c1-173d-4db7-9986-2873c3280e2a/barbican-worker/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.207558 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5684df76ff-6tw89_b1c9d4c1-173d-4db7-9986-2873c3280e2a/barbican-worker-log/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.442983 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9_46b05e85-6b4c-4aae-ab6b-39e026f48d49/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.453862 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/ceilometer-central-agent/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.566546 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/ceilometer-notification-agent/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.638068 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/sg-core/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.647265 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/proxy-httpd/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.852026 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c21557b2-c3ad-497a-ad42-ca81fff434bf/cinder-api/0.log" Dec 06 06:45:32 crc kubenswrapper[4957]: I1206 06:45:32.856160 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c21557b2-c3ad-497a-ad42-ca81fff434bf/cinder-api-log/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.062155 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_78ea9b70-3dde-4183-9ce4-c61326745bf2/cinder-scheduler/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.100844 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_78ea9b70-3dde-4183-9ce4-c61326745bf2/probe/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.149758 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5_a432bf6f-c43f-4b87-af4b-2af6281f1c25/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.289044 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc_2cb8adc2-7a82-4969-93d3-abac49e01d24/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.320572 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-7vzqk_3bac0712-aa19-4f2f-9645-eecdf3d19c5f/init/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.564910 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-7vzqk_3bac0712-aa19-4f2f-9645-eecdf3d19c5f/init/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.607550 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-shvb7_2dc93721-02c2-4553-960f-71955c73b548/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.635803 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-7vzqk_3bac0712-aa19-4f2f-9645-eecdf3d19c5f/dnsmasq-dns/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.778336 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2319a013-4227-4643-9b7f-d60b4d501af7/glance-httpd/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.827400 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2319a013-4227-4643-9b7f-d60b4d501af7/glance-log/0.log" Dec 06 06:45:33 crc kubenswrapper[4957]: I1206 06:45:33.991731 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e8deee48-5aef-4889-b7ab-b7f23144cbc8/glance-httpd/0.log" Dec 06 06:45:34 crc kubenswrapper[4957]: I1206 06:45:34.025162 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e8deee48-5aef-4889-b7ab-b7f23144cbc8/glance-log/0.log" Dec 06 06:45:34 crc kubenswrapper[4957]: I1206 06:45:34.204545 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c76d956-bsr6l_cb06a82d-1610-4985-8168-c95f32ad2ff6/horizon/0.log" Dec 06 06:45:34 crc kubenswrapper[4957]: I1206 06:45:34.385820 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz_26574c15-db39-470e-b739-cf03859b46aa/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:34 crc kubenswrapper[4957]: I1206 06:45:34.513458 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c76d956-bsr6l_cb06a82d-1610-4985-8168-c95f32ad2ff6/horizon-log/0.log" Dec 06 06:45:34 crc kubenswrapper[4957]: I1206 06:45:34.585440 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-2n487_87515d38-f8c8-4dd5-9373-fd2bd9ef7992/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:34 crc kubenswrapper[4957]: I1206 06:45:34.982399 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6a476151-cb3f-4502-8b9a-be049c359804/kube-state-metrics/0.log" Dec 06 06:45:34 crc kubenswrapper[4957]: I1206 06:45:34.999452 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7c7fdbc8b-njnq6_95511bb1-a0e3-4761-aba1-fd10b37df59f/keystone-api/0.log" Dec 06 06:45:35 crc kubenswrapper[4957]: I1206 06:45:35.187953 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz_5df54304-6d88-4027-ae21-40aaa328470f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:35 crc kubenswrapper[4957]: I1206 06:45:35.522812 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-575d75c97c-wblw4_a7c5d2f8-fce6-4689-8090-9ccbc8bb795e/neutron-api/0.log" Dec 06 06:45:35 crc kubenswrapper[4957]: I1206 06:45:35.567626 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-575d75c97c-wblw4_a7c5d2f8-fce6-4689-8090-9ccbc8bb795e/neutron-httpd/0.log" Dec 06 06:45:35 crc kubenswrapper[4957]: I1206 06:45:35.746681 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n_a0ec714a-36ad-46f4-aeca-a912266812f2/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:36 crc kubenswrapper[4957]: I1206 06:45:36.305108 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_9ce84760-5e40-4ea1-8bb2-81dc593c757a/nova-cell0-conductor-conductor/0.log" Dec 06 06:45:36 crc kubenswrapper[4957]: I1206 06:45:36.311681 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1ae0f500-0ce2-4374-afe8-cc65e465ac26/nova-api-log/0.log" Dec 06 06:45:36 crc kubenswrapper[4957]: I1206 06:45:36.415995 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1ae0f500-0ce2-4374-afe8-cc65e465ac26/nova-api-api/0.log" Dec 06 06:45:36 crc kubenswrapper[4957]: I1206 06:45:36.648161 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_6d950863-c9e9-4bc9-ae48-c8d1e63ad75c/nova-cell1-conductor-conductor/0.log" Dec 06 06:45:36 crc kubenswrapper[4957]: I1206 06:45:36.676247 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ecb88cbc-d465-49e9-9aaf-31aa46e18ae2/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 06:45:36 crc kubenswrapper[4957]: I1206 06:45:36.937258 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8dt69_312ec0f6-c580-4326-ba1a-b80706d87299/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:37 crc kubenswrapper[4957]: I1206 06:45:37.091439 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_67d6edc5-c0f1-4466-8b5b-99cd3bc06d17/nova-metadata-log/0.log" Dec 06 06:45:37 crc kubenswrapper[4957]: I1206 06:45:37.366041 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c683a5a8-170a-4880-b9fa-7b9635ce0751/nova-scheduler-scheduler/0.log" Dec 06 06:45:37 crc kubenswrapper[4957]: I1206 06:45:37.860324 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c10af3da-4556-4f2f-bfb8-886a6a93bfc4/mysql-bootstrap/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.100907 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c10af3da-4556-4f2f-bfb8-886a6a93bfc4/galera/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.140205 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c10af3da-4556-4f2f-bfb8-886a6a93bfc4/mysql-bootstrap/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.184459 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_67d6edc5-c0f1-4466-8b5b-99cd3bc06d17/nova-metadata-metadata/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.294546 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d87bf6f-a3d1-41ce-ba76-ed55aec8645e/mysql-bootstrap/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.553781 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d87bf6f-a3d1-41ce-ba76-ed55aec8645e/mysql-bootstrap/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.625851 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d87bf6f-a3d1-41ce-ba76-ed55aec8645e/galera/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.635936 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_663bc0fb-e06b-4c91-8a36-a97d32fa63f3/openstackclient/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.826011 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-q7r5j_e5adfb1c-2a78-42af-af7e-419bf4d73bd4/openstack-network-exporter/0.log" Dec 06 06:45:38 crc kubenswrapper[4957]: I1206 06:45:38.846104 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovsdb-server-init/0.log" Dec 06 06:45:39 crc kubenswrapper[4957]: I1206 06:45:39.732227 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovsdb-server-init/0.log" Dec 06 06:45:39 crc kubenswrapper[4957]: I1206 06:45:39.796701 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovs-vswitchd/0.log" Dec 06 06:45:39 crc kubenswrapper[4957]: I1206 06:45:39.803794 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovsdb-server/0.log" Dec 06 06:45:39 crc kubenswrapper[4957]: I1206 06:45:39.976853 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pmd7t_bf193218-2a15-4cf5-8ba0-59dc939445ff/ovn-controller/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.073315 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8bwfh_1eaa20e5-44dd-418a-94c2-fac0a16ef54a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.233848 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7262d5d8-b26d-4871-a3c7-9ff1f2c4009f/openstack-network-exporter/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.286953 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7262d5d8-b26d-4871-a3c7-9ff1f2c4009f/ovn-northd/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.429875 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d6ec6d60-cadc-488c-bb21-514a7acbfc70/openstack-network-exporter/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.456985 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d6ec6d60-cadc-488c-bb21-514a7acbfc70/ovsdbserver-nb/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.546756 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6c710a2d-a836-4154-9b55-70e3d6ad61a1/openstack-network-exporter/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.651586 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6c710a2d-a836-4154-9b55-70e3d6ad61a1/ovsdbserver-sb/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.784703 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69dd5567fb-lm2jl_8cada75b-b0d4-41c2-8d1f-3f146e28651d/placement-api/0.log" Dec 06 06:45:40 crc kubenswrapper[4957]: I1206 06:45:40.916794 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69dd5567fb-lm2jl_8cada75b-b0d4-41c2-8d1f-3f146e28651d/placement-log/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.020437 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a28833e8-d161-4dd0-816e-b476ab8d473b/setup-container/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.246824 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a28833e8-d161-4dd0-816e-b476ab8d473b/setup-container/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.314921 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a28833e8-d161-4dd0-816e-b476ab8d473b/rabbitmq/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.344387 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8504457c-7804-41fc-ac00-c2583a75137b/setup-container/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.522376 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8504457c-7804-41fc-ac00-c2583a75137b/rabbitmq/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.553293 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8504457c-7804-41fc-ac00-c2583a75137b/setup-container/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.605114 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv_101a78c1-076b-41be-a2bf-04f2f987924b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.832856 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jt982_a75e174a-7ba2-44d3-9d5d-060259173a48/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:41 crc kubenswrapper[4957]: I1206 06:45:41.892578 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-rc87w_6206b723-2b44-49ab-a3ec-93ce9d406ae8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.102658 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-rk8hd_9ca55297-0691-458e-85c2-a408a031c485/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.159176 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-c7hml_33b5c66b-1a29-44b1-9863-08dc69965dac/ssh-known-hosts-edpm-deployment/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.345869 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5789568449-4t258_475d23a9-0dda-4ba5-8af9-108574997a4f/proxy-server/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.414986 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5789568449-4t258_475d23a9-0dda-4ba5-8af9-108574997a4f/proxy-httpd/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.568380 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wm72g_43f9d83d-effb-4bc6-8550-6e4c32adaf8d/swift-ring-rebalance/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.632021 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-auditor/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.709378 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-reaper/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.778865 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-replicator/0.log" Dec 06 06:45:42 crc kubenswrapper[4957]: I1206 06:45:42.869145 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-server/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.001591 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-auditor/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.013106 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-replicator/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.147736 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-server/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.307405 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-updater/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.308525 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-auditor/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.346372 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-expirer/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.354806 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-replicator/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.488101 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-updater/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.527015 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-server/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.544103 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/swift-recon-cron/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.567377 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/rsync/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.812844 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_469a494f-1ceb-4d9f-bfe5-e0856e88f5bd/tempest-tests-tempest-tests-runner/0.log" Dec 06 06:45:43 crc kubenswrapper[4957]: I1206 06:45:43.832121 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g78cd_37341858-eaac-4c32-a5cc-7b06843f4fc2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:44 crc kubenswrapper[4957]: I1206 06:45:44.001242 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d42a9079-5c1e-424a-b79b-d469afb272c3/test-operator-logs-container/0.log" Dec 06 06:45:44 crc kubenswrapper[4957]: I1206 06:45:44.062315 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-f799z_7f74c7c3-a665-4958-aaad-15ed4e962fd0/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:45:45 crc kubenswrapper[4957]: I1206 06:45:45.918603 4957 scope.go:117] "RemoveContainer" containerID="e9fc2079c75ee4eeed9992d206edac0242fdb11945be37a78d6b98e4e4884ebd" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.217739 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_528a2b84-7047-444d-997e-4a92e0b5dbf9/memcached/0.log" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.780639 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qx5nh"] Dec 06 06:45:52 crc kubenswrapper[4957]: E1206 06:45:52.781083 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71d7471-1826-4dab-8b23-ce3ab05a5e75" containerName="container-00" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.781100 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71d7471-1826-4dab-8b23-ce3ab05a5e75" containerName="container-00" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.781312 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="c71d7471-1826-4dab-8b23-ce3ab05a5e75" containerName="container-00" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.782888 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.800469 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qx5nh"] Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.878195 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-utilities\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.878254 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbj4\" (UniqueName: \"kubernetes.io/projected/17ef80e9-4352-49bd-a153-7d03ab69aa93-kube-api-access-4fbj4\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.878283 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-catalog-content\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.980582 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-utilities\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.980642 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fbj4\" (UniqueName: \"kubernetes.io/projected/17ef80e9-4352-49bd-a153-7d03ab69aa93-kube-api-access-4fbj4\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.980687 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-catalog-content\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.981167 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-utilities\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:52 crc kubenswrapper[4957]: I1206 06:45:52.981251 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-catalog-content\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:53 crc kubenswrapper[4957]: I1206 06:45:53.379870 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fbj4\" (UniqueName: \"kubernetes.io/projected/17ef80e9-4352-49bd-a153-7d03ab69aa93-kube-api-access-4fbj4\") pod \"redhat-operators-qx5nh\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:53 crc kubenswrapper[4957]: I1206 06:45:53.405670 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:45:53 crc kubenswrapper[4957]: I1206 06:45:53.995562 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qx5nh"] Dec 06 06:45:54 crc kubenswrapper[4957]: I1206 06:45:54.866899 4957 generic.go:334] "Generic (PLEG): container finished" podID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerID="bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3" exitCode=0 Dec 06 06:45:54 crc kubenswrapper[4957]: I1206 06:45:54.867106 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qx5nh" event={"ID":"17ef80e9-4352-49bd-a153-7d03ab69aa93","Type":"ContainerDied","Data":"bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3"} Dec 06 06:45:54 crc kubenswrapper[4957]: I1206 06:45:54.867244 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qx5nh" event={"ID":"17ef80e9-4352-49bd-a153-7d03ab69aa93","Type":"ContainerStarted","Data":"0b0b32c2b98fffb0c6af2ed3f250423c5a33f37f0ed4f297d57a65f8177447a5"} Dec 06 06:45:55 crc kubenswrapper[4957]: I1206 06:45:55.877666 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qx5nh" event={"ID":"17ef80e9-4352-49bd-a153-7d03ab69aa93","Type":"ContainerStarted","Data":"0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3"} Dec 06 06:45:57 crc kubenswrapper[4957]: I1206 06:45:57.897634 4957 generic.go:334] "Generic (PLEG): container finished" podID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerID="0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3" exitCode=0 Dec 06 06:45:57 crc kubenswrapper[4957]: I1206 06:45:57.897784 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qx5nh" event={"ID":"17ef80e9-4352-49bd-a153-7d03ab69aa93","Type":"ContainerDied","Data":"0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3"} Dec 06 06:45:59 crc kubenswrapper[4957]: I1206 06:45:59.919392 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qx5nh" event={"ID":"17ef80e9-4352-49bd-a153-7d03ab69aa93","Type":"ContainerStarted","Data":"def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584"} Dec 06 06:45:59 crc kubenswrapper[4957]: I1206 06:45:59.947258 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qx5nh" podStartSLOduration=3.960476132 podStartE2EDuration="7.947236337s" podCreationTimestamp="2025-12-06 06:45:52 +0000 UTC" firstStartedPulling="2025-12-06 06:45:54.869042784 +0000 UTC m=+3994.519310416" lastFinishedPulling="2025-12-06 06:45:58.855802979 +0000 UTC m=+3998.506070621" observedRunningTime="2025-12-06 06:45:59.937208095 +0000 UTC m=+3999.587475727" watchObservedRunningTime="2025-12-06 06:45:59.947236337 +0000 UTC m=+3999.597503969" Dec 06 06:46:03 crc kubenswrapper[4957]: I1206 06:46:03.406480 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:46:03 crc kubenswrapper[4957]: I1206 06:46:03.407875 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:46:04 crc kubenswrapper[4957]: I1206 06:46:04.461641 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qx5nh" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="registry-server" probeResult="failure" output=< Dec 06 06:46:04 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 06:46:04 crc kubenswrapper[4957]: > Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.116722 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rpf52_6276ae06-7d72-409a-81f2-70ba70bf3d86/kube-rbac-proxy/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.160311 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rpf52_6276ae06-7d72-409a-81f2-70ba70bf3d86/manager/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.420536 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mzk64_bf2466dc-9cd3-4be7-9d17-846eccda13a8/kube-rbac-proxy/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.460869 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.519056 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.534102 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mzk64_bf2466dc-9cd3-4be7-9d17-846eccda13a8/manager/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.543140 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-9m6rj_e7ea0507-7466-4c88-a20b-2c936c6cdccb/kube-rbac-proxy/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.637727 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-9m6rj_e7ea0507-7466-4c88-a20b-2c936c6cdccb/manager/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.695045 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qx5nh"] Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.720045 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/util/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.929122 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/util/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.951484 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/pull/0.log" Dec 06 06:46:13 crc kubenswrapper[4957]: I1206 06:46:13.962684 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/pull/0.log" Dec 06 06:46:14 crc kubenswrapper[4957]: I1206 06:46:14.620484 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/util/0.log" Dec 06 06:46:14 crc kubenswrapper[4957]: I1206 06:46:14.755894 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/extract/0.log" Dec 06 06:46:14 crc kubenswrapper[4957]: I1206 06:46:14.769424 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/pull/0.log" Dec 06 06:46:14 crc kubenswrapper[4957]: I1206 06:46:14.825455 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d7r5m_c5d6f879-3842-4481-88bd-bb090eecd9fc/kube-rbac-proxy/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.014121 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rs7b8_ce46ec44-fc83-46e0-9f85-159e05452f0f/kube-rbac-proxy/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.037943 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qx5nh" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="registry-server" containerID="cri-o://def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584" gracePeriod=2 Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.038527 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d7r5m_c5d6f879-3842-4481-88bd-bb090eecd9fc/manager/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.096937 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rs7b8_ce46ec44-fc83-46e0-9f85-159e05452f0f/manager/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.209565 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-88lgf_5904f94f-274a-4b7f-92be-5b8acc5736ab/kube-rbac-proxy/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.281332 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-88lgf_5904f94f-274a-4b7f-92be-5b8acc5736ab/manager/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.391209 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9bc586d8b-5f5p9_e956383f-8ae1-444a-b876-a85b2bcd1bcb/kube-rbac-proxy/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.539372 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7nwnc_4affc2fc-c552-4a1b-b59e-31c7ff2cbb87/kube-rbac-proxy/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.636345 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9bc586d8b-5f5p9_e956383f-8ae1-444a-b876-a85b2bcd1bcb/manager/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.641106 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7nwnc_4affc2fc-c552-4a1b-b59e-31c7ff2cbb87/manager/0.log" Dec 06 06:46:15 crc kubenswrapper[4957]: I1206 06:46:15.791014 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7wgp_7aa1c48d-537b-4674-a284-1d929fc9cd16/kube-rbac-proxy/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.012081 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7wgp_7aa1c48d-537b-4674-a284-1d929fc9cd16/manager/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.014499 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-8mtps_b87d0440-b095-4540-a0a4-cb71461be4bd/kube-rbac-proxy/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.049825 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.049825 4957 generic.go:334] "Generic (PLEG): container finished" podID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerID="def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584" exitCode=0 Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.049886 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qx5nh" event={"ID":"17ef80e9-4352-49bd-a153-7d03ab69aa93","Type":"ContainerDied","Data":"def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584"} Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.050044 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qx5nh" event={"ID":"17ef80e9-4352-49bd-a153-7d03ab69aa93","Type":"ContainerDied","Data":"0b0b32c2b98fffb0c6af2ed3f250423c5a33f37f0ed4f297d57a65f8177447a5"} Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.050073 4957 scope.go:117] "RemoveContainer" containerID="def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.094301 4957 scope.go:117] "RemoveContainer" containerID="0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.117337 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-8mtps_b87d0440-b095-4540-a0a4-cb71461be4bd/manager/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.119945 4957 scope.go:117] "RemoveContainer" containerID="bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.162466 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-utilities\") pod \"17ef80e9-4352-49bd-a153-7d03ab69aa93\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.162564 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-catalog-content\") pod \"17ef80e9-4352-49bd-a153-7d03ab69aa93\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.162668 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fbj4\" (UniqueName: \"kubernetes.io/projected/17ef80e9-4352-49bd-a153-7d03ab69aa93-kube-api-access-4fbj4\") pod \"17ef80e9-4352-49bd-a153-7d03ab69aa93\" (UID: \"17ef80e9-4352-49bd-a153-7d03ab69aa93\") " Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.163945 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-utilities" (OuterVolumeSpecName: "utilities") pod "17ef80e9-4352-49bd-a153-7d03ab69aa93" (UID: "17ef80e9-4352-49bd-a153-7d03ab69aa93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.164569 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.181029 4957 scope.go:117] "RemoveContainer" containerID="def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.181261 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ef80e9-4352-49bd-a153-7d03ab69aa93-kube-api-access-4fbj4" (OuterVolumeSpecName: "kube-api-access-4fbj4") pod "17ef80e9-4352-49bd-a153-7d03ab69aa93" (UID: "17ef80e9-4352-49bd-a153-7d03ab69aa93"). InnerVolumeSpecName "kube-api-access-4fbj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:46:16 crc kubenswrapper[4957]: E1206 06:46:16.182035 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584\": container with ID starting with def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584 not found: ID does not exist" containerID="def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.182065 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584"} err="failed to get container status \"def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584\": rpc error: code = NotFound desc = could not find container \"def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584\": container with ID starting with def9f0893842766e15ea031f1688377452d5d45e63c1029a041f31babdba1584 not found: ID does not exist" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.182084 4957 scope.go:117] "RemoveContainer" containerID="0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3" Dec 06 06:46:16 crc kubenswrapper[4957]: E1206 06:46:16.182917 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3\": container with ID starting with 0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3 not found: ID does not exist" containerID="0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.182939 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3"} err="failed to get container status \"0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3\": rpc error: code = NotFound desc = could not find container \"0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3\": container with ID starting with 0cada371688db1a517274f7d8ac39cecc4836b918eb31d2ac0663e3a4fba75a3 not found: ID does not exist" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.182952 4957 scope.go:117] "RemoveContainer" containerID="bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3" Dec 06 06:46:16 crc kubenswrapper[4957]: E1206 06:46:16.184151 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3\": container with ID starting with bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3 not found: ID does not exist" containerID="bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.184176 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3"} err="failed to get container status \"bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3\": rpc error: code = NotFound desc = could not find container \"bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3\": container with ID starting with bfb86aeac0c2f79d7ab4c3612cd204fd521bc615eb4747a6bf6fd5ae1f732bd3 not found: ID does not exist" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.266401 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fbj4\" (UniqueName: \"kubernetes.io/projected/17ef80e9-4352-49bd-a153-7d03ab69aa93-kube-api-access-4fbj4\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.299824 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2x6nl_47c9992f-497a-4d57-8c37-c344ad49c199/kube-rbac-proxy/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.304625 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17ef80e9-4352-49bd-a153-7d03ab69aa93" (UID: "17ef80e9-4352-49bd-a153-7d03ab69aa93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.305557 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2x6nl_47c9992f-497a-4d57-8c37-c344ad49c199/manager/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.368123 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ef80e9-4352-49bd-a153-7d03ab69aa93-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.501192 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2cx79_4de58666-a4e1-4cae-99d7-2088192c51db/kube-rbac-proxy/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.526807 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2cx79_4de58666-a4e1-4cae-99d7-2088192c51db/manager/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.610876 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-rrq65_33674637-d0b9-40c9-a482-f4f18c39a80d/kube-rbac-proxy/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.737030 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-rrq65_33674637-d0b9-40c9-a482-f4f18c39a80d/manager/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.768626 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c2t2c_db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5/kube-rbac-proxy/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.786338 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c2t2c_db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5/manager/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.927716 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8_f9a79638-2a84-4bc8-a3c8-3289f3a1adae/manager/0.log" Dec 06 06:46:16 crc kubenswrapper[4957]: I1206 06:46:16.928278 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8_f9a79638-2a84-4bc8-a3c8-3289f3a1adae/kube-rbac-proxy/0.log" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.058053 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qx5nh" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.086249 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qx5nh"] Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.097239 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qx5nh"] Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.221482 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dr5bc_931598ec-e32e-4be7-81cf-1a5993ad1226/registry-server/0.log" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.312644 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-854f9c554b-shlk7_87dfe427-cd07-464f-9d50-ca308f553d91/operator/0.log" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.443447 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mc7wf_0a9ebb33-9805-4dec-a879-df7417c8ce7b/kube-rbac-proxy/0.log" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.476964 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mc7wf_0a9ebb33-9805-4dec-a879-df7417c8ce7b/manager/0.log" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.562651 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-g6zcq_d349e041-2168-4b8a-b6c4-a5840de8c9d8/kube-rbac-proxy/0.log" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.780292 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-6gccn_31332329-9dd4-4665-b1d0-9662ee3d1b8e/operator/0.log" Dec 06 06:46:17 crc kubenswrapper[4957]: I1206 06:46:17.781423 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-g6zcq_d349e041-2168-4b8a-b6c4-a5840de8c9d8/manager/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.023386 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xjbq2_2d7a7ff4-095a-4dff-9c96-25d03888a51e/kube-rbac-proxy/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.028890 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nr7jl_f2ab1e95-3e28-4d99-9340-f1eb24bc6dca/kube-rbac-proxy/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.062138 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xjbq2_2d7a7ff4-095a-4dff-9c96-25d03888a51e/manager/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.203077 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-75c7c544f9-bnr4h_b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e/manager/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.243975 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fvj4l_ea50a02a-23dc-4fcf-bcfd-048b66ce8152/kube-rbac-proxy/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.313564 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nr7jl_f2ab1e95-3e28-4d99-9340-f1eb24bc6dca/manager/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.355226 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fvj4l_ea50a02a-23dc-4fcf-bcfd-048b66ce8152/manager/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.426254 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k4mm8_ee5ec7b9-749f-4b24-826c-bebb580f7c5b/kube-rbac-proxy/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.494052 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k4mm8_ee5ec7b9-749f-4b24-826c-bebb580f7c5b/manager/0.log" Dec 06 06:46:18 crc kubenswrapper[4957]: I1206 06:46:18.674109 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" path="/var/lib/kubelet/pods/17ef80e9-4352-49bd-a153-7d03ab69aa93/volumes" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.011339 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7zqgc"] Dec 06 06:46:34 crc kubenswrapper[4957]: E1206 06:46:34.012219 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="extract-utilities" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.012232 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="extract-utilities" Dec 06 06:46:34 crc kubenswrapper[4957]: E1206 06:46:34.012261 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="registry-server" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.012276 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="registry-server" Dec 06 06:46:34 crc kubenswrapper[4957]: E1206 06:46:34.012285 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="extract-content" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.012292 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="extract-content" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.012471 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ef80e9-4352-49bd-a153-7d03ab69aa93" containerName="registry-server" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.013809 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.027847 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zqgc"] Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.087020 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7mgg\" (UniqueName: \"kubernetes.io/projected/74b0385a-72e7-47f5-b1c7-4d3c8b113881-kube-api-access-z7mgg\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.087376 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-catalog-content\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.087649 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-utilities\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.188937 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-utilities\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.189011 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7mgg\" (UniqueName: \"kubernetes.io/projected/74b0385a-72e7-47f5-b1c7-4d3c8b113881-kube-api-access-z7mgg\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.189104 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-catalog-content\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.189398 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-utilities\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.189575 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-catalog-content\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.216020 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7mgg\" (UniqueName: \"kubernetes.io/projected/74b0385a-72e7-47f5-b1c7-4d3c8b113881-kube-api-access-z7mgg\") pod \"community-operators-7zqgc\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.339272 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:34 crc kubenswrapper[4957]: I1206 06:46:34.828006 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zqgc"] Dec 06 06:46:34 crc kubenswrapper[4957]: W1206 06:46:34.838987 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74b0385a_72e7_47f5_b1c7_4d3c8b113881.slice/crio-f748079af3f10ac5e959156a06a9f1771db8042b87e9bf02783aaab0979b0488 WatchSource:0}: Error finding container f748079af3f10ac5e959156a06a9f1771db8042b87e9bf02783aaab0979b0488: Status 404 returned error can't find the container with id f748079af3f10ac5e959156a06a9f1771db8042b87e9bf02783aaab0979b0488 Dec 06 06:46:35 crc kubenswrapper[4957]: I1206 06:46:35.232790 4957 generic.go:334] "Generic (PLEG): container finished" podID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerID="6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52" exitCode=0 Dec 06 06:46:35 crc kubenswrapper[4957]: I1206 06:46:35.232898 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zqgc" event={"ID":"74b0385a-72e7-47f5-b1c7-4d3c8b113881","Type":"ContainerDied","Data":"6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52"} Dec 06 06:46:35 crc kubenswrapper[4957]: I1206 06:46:35.233135 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zqgc" event={"ID":"74b0385a-72e7-47f5-b1c7-4d3c8b113881","Type":"ContainerStarted","Data":"f748079af3f10ac5e959156a06a9f1771db8042b87e9bf02783aaab0979b0488"} Dec 06 06:46:36 crc kubenswrapper[4957]: I1206 06:46:36.245985 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zqgc" event={"ID":"74b0385a-72e7-47f5-b1c7-4d3c8b113881","Type":"ContainerStarted","Data":"518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1"} Dec 06 06:46:37 crc kubenswrapper[4957]: I1206 06:46:37.256865 4957 generic.go:334] "Generic (PLEG): container finished" podID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerID="518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1" exitCode=0 Dec 06 06:46:37 crc kubenswrapper[4957]: I1206 06:46:37.256968 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zqgc" event={"ID":"74b0385a-72e7-47f5-b1c7-4d3c8b113881","Type":"ContainerDied","Data":"518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1"} Dec 06 06:46:37 crc kubenswrapper[4957]: I1206 06:46:37.377682 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-m7xn8_92ffc396-7acd-4dd1-8bff-61901a7e6159/control-plane-machine-set-operator/0.log" Dec 06 06:46:37 crc kubenswrapper[4957]: I1206 06:46:37.509651 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5ld8_b9ad5cbf-adfe-4012-8968-0d7748db277d/kube-rbac-proxy/0.log" Dec 06 06:46:37 crc kubenswrapper[4957]: I1206 06:46:37.582663 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5ld8_b9ad5cbf-adfe-4012-8968-0d7748db277d/machine-api-operator/0.log" Dec 06 06:46:38 crc kubenswrapper[4957]: I1206 06:46:38.268065 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zqgc" event={"ID":"74b0385a-72e7-47f5-b1c7-4d3c8b113881","Type":"ContainerStarted","Data":"59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179"} Dec 06 06:46:38 crc kubenswrapper[4957]: I1206 06:46:38.304570 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7zqgc" podStartSLOduration=2.91979833 podStartE2EDuration="5.304549201s" podCreationTimestamp="2025-12-06 06:46:33 +0000 UTC" firstStartedPulling="2025-12-06 06:46:35.234711292 +0000 UTC m=+4034.884978934" lastFinishedPulling="2025-12-06 06:46:37.619462163 +0000 UTC m=+4037.269729805" observedRunningTime="2025-12-06 06:46:38.29347433 +0000 UTC m=+4037.943741992" watchObservedRunningTime="2025-12-06 06:46:38.304549201 +0000 UTC m=+4037.954816853" Dec 06 06:46:44 crc kubenswrapper[4957]: I1206 06:46:44.339731 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:44 crc kubenswrapper[4957]: I1206 06:46:44.340433 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:45 crc kubenswrapper[4957]: I1206 06:46:45.019079 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:45 crc kubenswrapper[4957]: I1206 06:46:45.382111 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:45 crc kubenswrapper[4957]: I1206 06:46:45.449307 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zqgc"] Dec 06 06:46:47 crc kubenswrapper[4957]: I1206 06:46:47.349627 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7zqgc" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="registry-server" containerID="cri-o://59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179" gracePeriod=2 Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.342178 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.358725 4957 generic.go:334] "Generic (PLEG): container finished" podID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerID="59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179" exitCode=0 Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.358760 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zqgc" event={"ID":"74b0385a-72e7-47f5-b1c7-4d3c8b113881","Type":"ContainerDied","Data":"59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179"} Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.358802 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zqgc" event={"ID":"74b0385a-72e7-47f5-b1c7-4d3c8b113881","Type":"ContainerDied","Data":"f748079af3f10ac5e959156a06a9f1771db8042b87e9bf02783aaab0979b0488"} Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.358807 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zqgc" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.358821 4957 scope.go:117] "RemoveContainer" containerID="59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.377131 4957 scope.go:117] "RemoveContainer" containerID="518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.407047 4957 scope.go:117] "RemoveContainer" containerID="6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.447457 4957 scope.go:117] "RemoveContainer" containerID="59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179" Dec 06 06:46:48 crc kubenswrapper[4957]: E1206 06:46:48.449573 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179\": container with ID starting with 59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179 not found: ID does not exist" containerID="59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.449652 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179"} err="failed to get container status \"59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179\": rpc error: code = NotFound desc = could not find container \"59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179\": container with ID starting with 59f5e166893fdb79ade66d5deaca07f48287efd3bd386a22862a00a7d67b6179 not found: ID does not exist" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.449680 4957 scope.go:117] "RemoveContainer" containerID="518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1" Dec 06 06:46:48 crc kubenswrapper[4957]: E1206 06:46:48.450307 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1\": container with ID starting with 518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1 not found: ID does not exist" containerID="518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.450355 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1"} err="failed to get container status \"518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1\": rpc error: code = NotFound desc = could not find container \"518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1\": container with ID starting with 518e0ec919bba18511061e5c200e78c4a0bf98945b480cb4d6f6a20a2597e7f1 not found: ID does not exist" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.450387 4957 scope.go:117] "RemoveContainer" containerID="6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52" Dec 06 06:46:48 crc kubenswrapper[4957]: E1206 06:46:48.450650 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52\": container with ID starting with 6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52 not found: ID does not exist" containerID="6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.450743 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52"} err="failed to get container status \"6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52\": rpc error: code = NotFound desc = could not find container \"6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52\": container with ID starting with 6d34575ff0b15c7b783a472fe27a55c302343e48896ffccd53e9aa0963565c52 not found: ID does not exist" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.509754 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-catalog-content\") pod \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.509862 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-utilities\") pod \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.509960 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7mgg\" (UniqueName: \"kubernetes.io/projected/74b0385a-72e7-47f5-b1c7-4d3c8b113881-kube-api-access-z7mgg\") pod \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\" (UID: \"74b0385a-72e7-47f5-b1c7-4d3c8b113881\") " Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.511358 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-utilities" (OuterVolumeSpecName: "utilities") pod "74b0385a-72e7-47f5-b1c7-4d3c8b113881" (UID: "74b0385a-72e7-47f5-b1c7-4d3c8b113881"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.514760 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b0385a-72e7-47f5-b1c7-4d3c8b113881-kube-api-access-z7mgg" (OuterVolumeSpecName: "kube-api-access-z7mgg") pod "74b0385a-72e7-47f5-b1c7-4d3c8b113881" (UID: "74b0385a-72e7-47f5-b1c7-4d3c8b113881"). InnerVolumeSpecName "kube-api-access-z7mgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.560295 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74b0385a-72e7-47f5-b1c7-4d3c8b113881" (UID: "74b0385a-72e7-47f5-b1c7-4d3c8b113881"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.612872 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.612908 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7mgg\" (UniqueName: \"kubernetes.io/projected/74b0385a-72e7-47f5-b1c7-4d3c8b113881-kube-api-access-z7mgg\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.612921 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b0385a-72e7-47f5-b1c7-4d3c8b113881-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.702373 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zqgc"] Dec 06 06:46:48 crc kubenswrapper[4957]: I1206 06:46:48.713954 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7zqgc"] Dec 06 06:46:50 crc kubenswrapper[4957]: I1206 06:46:50.675287 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" path="/var/lib/kubelet/pods/74b0385a-72e7-47f5-b1c7-4d3c8b113881/volumes" Dec 06 06:46:51 crc kubenswrapper[4957]: I1206 06:46:51.049909 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-59bbt_ff08ac61-0abb-4f45-a647-57aca5172e36/cert-manager-controller/0.log" Dec 06 06:46:51 crc kubenswrapper[4957]: I1206 06:46:51.198727 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-6557k_14f39aaa-5904-4a8d-86ad-ac487456788b/cert-manager-cainjector/0.log" Dec 06 06:46:51 crc kubenswrapper[4957]: I1206 06:46:51.255219 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-b256m_8ebad773-0957-4bb8-b583-a0cff091ff46/cert-manager-webhook/0.log" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.747238 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j4f7d"] Dec 06 06:47:01 crc kubenswrapper[4957]: E1206 06:47:01.748242 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="extract-utilities" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.748256 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="extract-utilities" Dec 06 06:47:01 crc kubenswrapper[4957]: E1206 06:47:01.748277 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="registry-server" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.748283 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="registry-server" Dec 06 06:47:01 crc kubenswrapper[4957]: E1206 06:47:01.748299 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="extract-content" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.748305 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="extract-content" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.748552 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b0385a-72e7-47f5-b1c7-4d3c8b113881" containerName="registry-server" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.750011 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.761781 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4f7d"] Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.815517 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-utilities\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.815624 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc9xd\" (UniqueName: \"kubernetes.io/projected/eeae6202-e05b-4752-b6be-9b018ecab58f-kube-api-access-jc9xd\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.815754 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-catalog-content\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.917163 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-catalog-content\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.917252 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-utilities\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.917325 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc9xd\" (UniqueName: \"kubernetes.io/projected/eeae6202-e05b-4752-b6be-9b018ecab58f-kube-api-access-jc9xd\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.917698 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-catalog-content\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.917798 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-utilities\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:01 crc kubenswrapper[4957]: I1206 06:47:01.952257 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc9xd\" (UniqueName: \"kubernetes.io/projected/eeae6202-e05b-4752-b6be-9b018ecab58f-kube-api-access-jc9xd\") pod \"redhat-marketplace-j4f7d\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:02 crc kubenswrapper[4957]: I1206 06:47:02.084779 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:02 crc kubenswrapper[4957]: I1206 06:47:02.648802 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4f7d"] Dec 06 06:47:02 crc kubenswrapper[4957]: W1206 06:47:02.655150 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeae6202_e05b_4752_b6be_9b018ecab58f.slice/crio-557a704d9000a1501fc7504152a3bdff97b2b1ba0f0a3809db243b0206f8016e WatchSource:0}: Error finding container 557a704d9000a1501fc7504152a3bdff97b2b1ba0f0a3809db243b0206f8016e: Status 404 returned error can't find the container with id 557a704d9000a1501fc7504152a3bdff97b2b1ba0f0a3809db243b0206f8016e Dec 06 06:47:03 crc kubenswrapper[4957]: I1206 06:47:03.509990 4957 generic.go:334] "Generic (PLEG): container finished" podID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerID="70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7" exitCode=0 Dec 06 06:47:03 crc kubenswrapper[4957]: I1206 06:47:03.510102 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4f7d" event={"ID":"eeae6202-e05b-4752-b6be-9b018ecab58f","Type":"ContainerDied","Data":"70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7"} Dec 06 06:47:03 crc kubenswrapper[4957]: I1206 06:47:03.510585 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4f7d" event={"ID":"eeae6202-e05b-4752-b6be-9b018ecab58f","Type":"ContainerStarted","Data":"557a704d9000a1501fc7504152a3bdff97b2b1ba0f0a3809db243b0206f8016e"} Dec 06 06:47:05 crc kubenswrapper[4957]: I1206 06:47:05.531758 4957 generic.go:334] "Generic (PLEG): container finished" podID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerID="e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9" exitCode=0 Dec 06 06:47:05 crc kubenswrapper[4957]: I1206 06:47:05.531900 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4f7d" event={"ID":"eeae6202-e05b-4752-b6be-9b018ecab58f","Type":"ContainerDied","Data":"e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9"} Dec 06 06:47:06 crc kubenswrapper[4957]: I1206 06:47:06.470289 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-wtc5f_cbf7952f-0cb5-4873-a116-b878ccdbf7e2/nmstate-console-plugin/0.log" Dec 06 06:47:06 crc kubenswrapper[4957]: I1206 06:47:06.544252 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4f7d" event={"ID":"eeae6202-e05b-4752-b6be-9b018ecab58f","Type":"ContainerStarted","Data":"f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8"} Dec 06 06:47:06 crc kubenswrapper[4957]: I1206 06:47:06.544306 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-x8lpv_211db7e0-522a-4635-a279-1d58502c1d1f/nmstate-handler/0.log" Dec 06 06:47:06 crc kubenswrapper[4957]: I1206 06:47:06.563417 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j4f7d" podStartSLOduration=3.115964067 podStartE2EDuration="5.563394668s" podCreationTimestamp="2025-12-06 06:47:01 +0000 UTC" firstStartedPulling="2025-12-06 06:47:03.513342535 +0000 UTC m=+4063.163610167" lastFinishedPulling="2025-12-06 06:47:05.960773136 +0000 UTC m=+4065.611040768" observedRunningTime="2025-12-06 06:47:06.560938092 +0000 UTC m=+4066.211205734" watchObservedRunningTime="2025-12-06 06:47:06.563394668 +0000 UTC m=+4066.213662300" Dec 06 06:47:06 crc kubenswrapper[4957]: I1206 06:47:06.784506 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mlsvn_a60b6144-6a1f-4a86-ae5e-84d5771a8c9c/nmstate-metrics/0.log" Dec 06 06:47:06 crc kubenswrapper[4957]: I1206 06:47:06.838278 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mlsvn_a60b6144-6a1f-4a86-ae5e-84d5771a8c9c/kube-rbac-proxy/0.log" Dec 06 06:47:07 crc kubenswrapper[4957]: I1206 06:47:07.052287 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-2d6j5_44a5fc06-b6e1-4852-a528-f92117991c7a/nmstate-operator/0.log" Dec 06 06:47:07 crc kubenswrapper[4957]: I1206 06:47:07.075332 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-6r8nd_f85d16c2-33be-4772-a597-fa4393b5846c/nmstate-webhook/0.log" Dec 06 06:47:12 crc kubenswrapper[4957]: I1206 06:47:12.085998 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:12 crc kubenswrapper[4957]: I1206 06:47:12.086372 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:12 crc kubenswrapper[4957]: I1206 06:47:12.134670 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:12 crc kubenswrapper[4957]: I1206 06:47:12.653602 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:12 crc kubenswrapper[4957]: I1206 06:47:12.728050 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4f7d"] Dec 06 06:47:14 crc kubenswrapper[4957]: I1206 06:47:14.606944 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j4f7d" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="registry-server" containerID="cri-o://f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8" gracePeriod=2 Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.169065 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.366463 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-utilities\") pod \"eeae6202-e05b-4752-b6be-9b018ecab58f\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.366599 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc9xd\" (UniqueName: \"kubernetes.io/projected/eeae6202-e05b-4752-b6be-9b018ecab58f-kube-api-access-jc9xd\") pod \"eeae6202-e05b-4752-b6be-9b018ecab58f\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.366743 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-catalog-content\") pod \"eeae6202-e05b-4752-b6be-9b018ecab58f\" (UID: \"eeae6202-e05b-4752-b6be-9b018ecab58f\") " Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.367459 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-utilities" (OuterVolumeSpecName: "utilities") pod "eeae6202-e05b-4752-b6be-9b018ecab58f" (UID: "eeae6202-e05b-4752-b6be-9b018ecab58f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.393516 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eeae6202-e05b-4752-b6be-9b018ecab58f" (UID: "eeae6202-e05b-4752-b6be-9b018ecab58f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.468927 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.468960 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeae6202-e05b-4752-b6be-9b018ecab58f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.619319 4957 generic.go:334] "Generic (PLEG): container finished" podID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerID="f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8" exitCode=0 Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.619388 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4f7d" event={"ID":"eeae6202-e05b-4752-b6be-9b018ecab58f","Type":"ContainerDied","Data":"f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8"} Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.619452 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4f7d" event={"ID":"eeae6202-e05b-4752-b6be-9b018ecab58f","Type":"ContainerDied","Data":"557a704d9000a1501fc7504152a3bdff97b2b1ba0f0a3809db243b0206f8016e"} Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.619475 4957 scope.go:117] "RemoveContainer" containerID="f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.619498 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4f7d" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.638076 4957 scope.go:117] "RemoveContainer" containerID="e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.784688 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeae6202-e05b-4752-b6be-9b018ecab58f-kube-api-access-jc9xd" (OuterVolumeSpecName: "kube-api-access-jc9xd") pod "eeae6202-e05b-4752-b6be-9b018ecab58f" (UID: "eeae6202-e05b-4752-b6be-9b018ecab58f"). InnerVolumeSpecName "kube-api-access-jc9xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.791982 4957 scope.go:117] "RemoveContainer" containerID="70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.875347 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc9xd\" (UniqueName: \"kubernetes.io/projected/eeae6202-e05b-4752-b6be-9b018ecab58f-kube-api-access-jc9xd\") on node \"crc\" DevicePath \"\"" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.930429 4957 scope.go:117] "RemoveContainer" containerID="f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8" Dec 06 06:47:15 crc kubenswrapper[4957]: E1206 06:47:15.937757 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8\": container with ID starting with f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8 not found: ID does not exist" containerID="f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.937858 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8"} err="failed to get container status \"f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8\": rpc error: code = NotFound desc = could not find container \"f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8\": container with ID starting with f409dce758ae6b0696ead4763bc1e93b6ce932da95c0df3803cb1c05ca5edfc8 not found: ID does not exist" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.937900 4957 scope.go:117] "RemoveContainer" containerID="e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9" Dec 06 06:47:15 crc kubenswrapper[4957]: E1206 06:47:15.939153 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9\": container with ID starting with e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9 not found: ID does not exist" containerID="e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.939198 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9"} err="failed to get container status \"e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9\": rpc error: code = NotFound desc = could not find container \"e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9\": container with ID starting with e846aa30922bc92ee6e94a138cf407242f3f05b6d4b7293b2364c8d210be02d9 not found: ID does not exist" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.939234 4957 scope.go:117] "RemoveContainer" containerID="70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7" Dec 06 06:47:15 crc kubenswrapper[4957]: E1206 06:47:15.941386 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7\": container with ID starting with 70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7 not found: ID does not exist" containerID="70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.941472 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7"} err="failed to get container status \"70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7\": rpc error: code = NotFound desc = could not find container \"70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7\": container with ID starting with 70815562f9c9e0cf2bbe455c183ae3aa0620cb71a08040d87d77db65f0e405a7 not found: ID does not exist" Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.974737 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4f7d"] Dec 06 06:47:15 crc kubenswrapper[4957]: I1206 06:47:15.986919 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4f7d"] Dec 06 06:47:16 crc kubenswrapper[4957]: I1206 06:47:16.672584 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" path="/var/lib/kubelet/pods/eeae6202-e05b-4752-b6be-9b018ecab58f/volumes" Dec 06 06:47:21 crc kubenswrapper[4957]: I1206 06:47:21.730656 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzfkk_5ee6dee4-31cd-4d22-905c-a0339c07f1b1/kube-rbac-proxy/0.log" Dec 06 06:47:21 crc kubenswrapper[4957]: I1206 06:47:21.862216 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzfkk_5ee6dee4-31cd-4d22-905c-a0339c07f1b1/controller/0.log" Dec 06 06:47:21 crc kubenswrapper[4957]: I1206 06:47:21.899797 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.091244 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.100356 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.127542 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.143726 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.302774 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.323049 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.325495 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.333684 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.481489 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.516234 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.518583 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.543047 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/controller/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.759347 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/kube-rbac-proxy/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.936656 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/kube-rbac-proxy-frr/0.log" Dec 06 06:47:22 crc kubenswrapper[4957]: I1206 06:47:22.951792 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/frr-metrics/0.log" Dec 06 06:47:23 crc kubenswrapper[4957]: I1206 06:47:23.073019 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/reloader/0.log" Dec 06 06:47:23 crc kubenswrapper[4957]: I1206 06:47:23.177757 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-sbt5p_5338fe04-633c-405b-8079-1001859b804f/frr-k8s-webhook-server/0.log" Dec 06 06:47:23 crc kubenswrapper[4957]: I1206 06:47:23.430131 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9b6b64df4-mn6tm_c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f/manager/0.log" Dec 06 06:47:23 crc kubenswrapper[4957]: I1206 06:47:23.566316 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5568f7b8bc-8wsfl_c3dd44fb-8340-44f9-af68-f7e628a59b6c/webhook-server/0.log" Dec 06 06:47:23 crc kubenswrapper[4957]: I1206 06:47:23.709889 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2jbdq_fa194652-baa4-424b-bcd0-736ac7295ec4/kube-rbac-proxy/0.log" Dec 06 06:47:23 crc kubenswrapper[4957]: I1206 06:47:23.833547 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/frr/0.log" Dec 06 06:47:24 crc kubenswrapper[4957]: I1206 06:47:24.194743 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2jbdq_fa194652-baa4-424b-bcd0-736ac7295ec4/speaker/0.log" Dec 06 06:47:36 crc kubenswrapper[4957]: I1206 06:47:36.428327 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/util/0.log" Dec 06 06:47:36 crc kubenswrapper[4957]: I1206 06:47:36.649387 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/util/0.log" Dec 06 06:47:36 crc kubenswrapper[4957]: I1206 06:47:36.702233 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/pull/0.log" Dec 06 06:47:36 crc kubenswrapper[4957]: I1206 06:47:36.716264 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/pull/0.log" Dec 06 06:47:36 crc kubenswrapper[4957]: I1206 06:47:36.874622 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/util/0.log" Dec 06 06:47:36 crc kubenswrapper[4957]: I1206 06:47:36.883485 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/pull/0.log" Dec 06 06:47:36 crc kubenswrapper[4957]: I1206 06:47:36.890758 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/extract/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.090223 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/util/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.279729 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/util/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.280083 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/pull/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.318202 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/pull/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.439309 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/util/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.486928 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/extract/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.654390 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-utilities/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.654920 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/pull/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.823630 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-utilities/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.865748 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-content/0.log" Dec 06 06:47:37 crc kubenswrapper[4957]: I1206 06:47:37.870099 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-content/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.038700 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-utilities/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.045201 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-content/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.266509 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-utilities/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.438116 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-utilities/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.467590 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-content/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.537024 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-content/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.608070 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/registry-server/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.726681 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-utilities/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.784644 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-content/0.log" Dec 06 06:47:38 crc kubenswrapper[4957]: I1206 06:47:38.959522 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cbbgh_de45cbbb-4070-4df4-8b0e-c6dbe32a1a33/marketplace-operator/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.026827 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-utilities/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.295955 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/registry-server/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.305303 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-content/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.346138 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-content/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.357379 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-utilities/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.490756 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-content/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.493140 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-utilities/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.648016 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/registry-server/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.707242 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-utilities/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.860409 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-utilities/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.864976 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-content/0.log" Dec 06 06:47:39 crc kubenswrapper[4957]: I1206 06:47:39.921655 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-content/0.log" Dec 06 06:47:40 crc kubenswrapper[4957]: I1206 06:47:40.134011 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-content/0.log" Dec 06 06:47:40 crc kubenswrapper[4957]: I1206 06:47:40.150597 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-utilities/0.log" Dec 06 06:47:40 crc kubenswrapper[4957]: I1206 06:47:40.180369 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:47:40 crc kubenswrapper[4957]: I1206 06:47:40.180425 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:47:40 crc kubenswrapper[4957]: I1206 06:47:40.627458 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/registry-server/0.log" Dec 06 06:48:10 crc kubenswrapper[4957]: I1206 06:48:10.193317 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:48:10 crc kubenswrapper[4957]: I1206 06:48:10.193854 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:48:13 crc kubenswrapper[4957]: E1206 06:48:13.706527 4957 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.151:52328->38.102.83.151:36993: write tcp 38.102.83.151:52328->38.102.83.151:36993: write: connection reset by peer Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.180691 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.181471 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.181516 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.182258 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ccddb8ad5bca662c4e56f40c21bfbea3eb92dbae28847228b123252634f9d79"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.182319 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://6ccddb8ad5bca662c4e56f40c21bfbea3eb92dbae28847228b123252634f9d79" gracePeriod=600 Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.383607 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="6ccddb8ad5bca662c4e56f40c21bfbea3eb92dbae28847228b123252634f9d79" exitCode=0 Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.383706 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"6ccddb8ad5bca662c4e56f40c21bfbea3eb92dbae28847228b123252634f9d79"} Dec 06 06:48:40 crc kubenswrapper[4957]: I1206 06:48:40.384013 4957 scope.go:117] "RemoveContainer" containerID="21f6d53407039cb44fcfab7d92676657e0efe964720385605c838365aad4e508" Dec 06 06:48:41 crc kubenswrapper[4957]: I1206 06:48:41.397519 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a"} Dec 06 06:49:21 crc kubenswrapper[4957]: I1206 06:49:21.822857 4957 generic.go:334] "Generic (PLEG): container finished" podID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerID="c29bfa67dfacfb218cff8778e68c95cb16e9bd8b46df8e07034fbcbd4865c741" exitCode=0 Dec 06 06:49:21 crc kubenswrapper[4957]: I1206 06:49:21.822929 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vlspc/must-gather-sbsvd" event={"ID":"a3a3017f-6621-4931-80cd-040dc7a81d9f","Type":"ContainerDied","Data":"c29bfa67dfacfb218cff8778e68c95cb16e9bd8b46df8e07034fbcbd4865c741"} Dec 06 06:49:21 crc kubenswrapper[4957]: I1206 06:49:21.824063 4957 scope.go:117] "RemoveContainer" containerID="c29bfa67dfacfb218cff8778e68c95cb16e9bd8b46df8e07034fbcbd4865c741" Dec 06 06:49:22 crc kubenswrapper[4957]: I1206 06:49:22.766216 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vlspc_must-gather-sbsvd_a3a3017f-6621-4931-80cd-040dc7a81d9f/gather/0.log" Dec 06 06:49:30 crc kubenswrapper[4957]: I1206 06:49:30.560584 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vlspc/must-gather-sbsvd"] Dec 06 06:49:30 crc kubenswrapper[4957]: I1206 06:49:30.561371 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-vlspc/must-gather-sbsvd" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerName="copy" containerID="cri-o://76ded26ea0e3694c4ad59de8bd22b9028413b3e29389b7eee8e441ab3272a63f" gracePeriod=2 Dec 06 06:49:30 crc kubenswrapper[4957]: I1206 06:49:30.569677 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vlspc/must-gather-sbsvd"] Dec 06 06:49:30 crc kubenswrapper[4957]: I1206 06:49:30.900562 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vlspc_must-gather-sbsvd_a3a3017f-6621-4931-80cd-040dc7a81d9f/copy/0.log" Dec 06 06:49:30 crc kubenswrapper[4957]: I1206 06:49:30.901182 4957 generic.go:334] "Generic (PLEG): container finished" podID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerID="76ded26ea0e3694c4ad59de8bd22b9028413b3e29389b7eee8e441ab3272a63f" exitCode=143 Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.406471 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vlspc_must-gather-sbsvd_a3a3017f-6621-4931-80cd-040dc7a81d9f/copy/0.log" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.407229 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.561049 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a3a3017f-6621-4931-80cd-040dc7a81d9f-must-gather-output\") pod \"a3a3017f-6621-4931-80cd-040dc7a81d9f\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.561209 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6bpw\" (UniqueName: \"kubernetes.io/projected/a3a3017f-6621-4931-80cd-040dc7a81d9f-kube-api-access-v6bpw\") pod \"a3a3017f-6621-4931-80cd-040dc7a81d9f\" (UID: \"a3a3017f-6621-4931-80cd-040dc7a81d9f\") " Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.567769 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a3017f-6621-4931-80cd-040dc7a81d9f-kube-api-access-v6bpw" (OuterVolumeSpecName: "kube-api-access-v6bpw") pod "a3a3017f-6621-4931-80cd-040dc7a81d9f" (UID: "a3a3017f-6621-4931-80cd-040dc7a81d9f"). InnerVolumeSpecName "kube-api-access-v6bpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.665902 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6bpw\" (UniqueName: \"kubernetes.io/projected/a3a3017f-6621-4931-80cd-040dc7a81d9f-kube-api-access-v6bpw\") on node \"crc\" DevicePath \"\"" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.751584 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3a3017f-6621-4931-80cd-040dc7a81d9f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a3a3017f-6621-4931-80cd-040dc7a81d9f" (UID: "a3a3017f-6621-4931-80cd-040dc7a81d9f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.768554 4957 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a3a3017f-6621-4931-80cd-040dc7a81d9f-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.912160 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vlspc_must-gather-sbsvd_a3a3017f-6621-4931-80cd-040dc7a81d9f/copy/0.log" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.912745 4957 scope.go:117] "RemoveContainer" containerID="76ded26ea0e3694c4ad59de8bd22b9028413b3e29389b7eee8e441ab3272a63f" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.912852 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vlspc/must-gather-sbsvd" Dec 06 06:49:31 crc kubenswrapper[4957]: I1206 06:49:31.932721 4957 scope.go:117] "RemoveContainer" containerID="c29bfa67dfacfb218cff8778e68c95cb16e9bd8b46df8e07034fbcbd4865c741" Dec 06 06:49:32 crc kubenswrapper[4957]: I1206 06:49:32.673046 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" path="/var/lib/kubelet/pods/a3a3017f-6621-4931-80cd-040dc7a81d9f/volumes" Dec 06 06:50:40 crc kubenswrapper[4957]: I1206 06:50:40.180922 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:50:40 crc kubenswrapper[4957]: I1206 06:50:40.181588 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:50:46 crc kubenswrapper[4957]: I1206 06:50:46.211680 4957 scope.go:117] "RemoveContainer" containerID="4a29b460acf8d10d348da4dba51b8ef85495afaa2f1638aaf3157d65dbb196f9" Dec 06 06:51:10 crc kubenswrapper[4957]: I1206 06:51:10.180616 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:51:10 crc kubenswrapper[4957]: I1206 06:51:10.181970 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:51:40 crc kubenswrapper[4957]: I1206 06:51:40.181091 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:51:40 crc kubenswrapper[4957]: I1206 06:51:40.181611 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:51:40 crc kubenswrapper[4957]: I1206 06:51:40.181659 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 06:51:40 crc kubenswrapper[4957]: I1206 06:51:40.182421 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 06:51:40 crc kubenswrapper[4957]: I1206 06:51:40.182493 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" gracePeriod=600 Dec 06 06:51:40 crc kubenswrapper[4957]: E1206 06:51:40.320304 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:51:41 crc kubenswrapper[4957]: I1206 06:51:41.215274 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" exitCode=0 Dec 06 06:51:41 crc kubenswrapper[4957]: I1206 06:51:41.215491 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a"} Dec 06 06:51:41 crc kubenswrapper[4957]: I1206 06:51:41.215688 4957 scope.go:117] "RemoveContainer" containerID="6ccddb8ad5bca662c4e56f40c21bfbea3eb92dbae28847228b123252634f9d79" Dec 06 06:51:41 crc kubenswrapper[4957]: I1206 06:51:41.216554 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:51:41 crc kubenswrapper[4957]: E1206 06:51:41.216908 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:51:52 crc kubenswrapper[4957]: I1206 06:51:52.662326 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:51:52 crc kubenswrapper[4957]: E1206 06:51:52.663150 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:52:04 crc kubenswrapper[4957]: I1206 06:52:04.662205 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:52:04 crc kubenswrapper[4957]: E1206 06:52:04.663145 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:52:16 crc kubenswrapper[4957]: I1206 06:52:16.662749 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:52:16 crc kubenswrapper[4957]: E1206 06:52:16.664501 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.434347 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9qkxf/must-gather-8sgxr"] Dec 06 06:52:28 crc kubenswrapper[4957]: E1206 06:52:28.435441 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerName="gather" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435459 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerName="gather" Dec 06 06:52:28 crc kubenswrapper[4957]: E1206 06:52:28.435483 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="extract-utilities" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435492 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="extract-utilities" Dec 06 06:52:28 crc kubenswrapper[4957]: E1206 06:52:28.435510 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="extract-content" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435519 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="extract-content" Dec 06 06:52:28 crc kubenswrapper[4957]: E1206 06:52:28.435534 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="registry-server" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435542 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="registry-server" Dec 06 06:52:28 crc kubenswrapper[4957]: E1206 06:52:28.435571 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerName="copy" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435579 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerName="copy" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435850 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeae6202-e05b-4752-b6be-9b018ecab58f" containerName="registry-server" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435876 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerName="gather" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.435896 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a3017f-6621-4931-80cd-040dc7a81d9f" containerName="copy" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.438591 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.442234 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9qkxf"/"openshift-service-ca.crt" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.442251 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9qkxf"/"default-dockercfg-9fdqs" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.442736 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9qkxf"/"kube-root-ca.crt" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.470870 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9qkxf/must-gather-8sgxr"] Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.500708 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-must-gather-output\") pod \"must-gather-8sgxr\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.501106 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnsq7\" (UniqueName: \"kubernetes.io/projected/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-kube-api-access-lnsq7\") pod \"must-gather-8sgxr\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.602675 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-must-gather-output\") pod \"must-gather-8sgxr\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.603088 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnsq7\" (UniqueName: \"kubernetes.io/projected/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-kube-api-access-lnsq7\") pod \"must-gather-8sgxr\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.603950 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-must-gather-output\") pod \"must-gather-8sgxr\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.621208 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnsq7\" (UniqueName: \"kubernetes.io/projected/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-kube-api-access-lnsq7\") pod \"must-gather-8sgxr\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.663077 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:52:28 crc kubenswrapper[4957]: E1206 06:52:28.663533 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:52:28 crc kubenswrapper[4957]: I1206 06:52:28.760569 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:52:29 crc kubenswrapper[4957]: I1206 06:52:29.228479 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9qkxf/must-gather-8sgxr"] Dec 06 06:52:30 crc kubenswrapper[4957]: I1206 06:52:30.694107 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" event={"ID":"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a","Type":"ContainerStarted","Data":"b8044f1232b7628997571c1e8f10ce4e153db9779f26dcb3dcaedb1ef12b8bbc"} Dec 06 06:52:30 crc kubenswrapper[4957]: I1206 06:52:30.694629 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" event={"ID":"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a","Type":"ContainerStarted","Data":"e100d4da4629ad227b749c1fee0c2c8df1a72baf374b059a17134b6569ebed57"} Dec 06 06:52:30 crc kubenswrapper[4957]: I1206 06:52:30.694645 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" event={"ID":"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a","Type":"ContainerStarted","Data":"c5ab52397c652f85fa1f462c9779e543ccf25e268f49b881bb8ba46a42498e8f"} Dec 06 06:52:30 crc kubenswrapper[4957]: I1206 06:52:30.712041 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" podStartSLOduration=2.712020469 podStartE2EDuration="2.712020469s" podCreationTimestamp="2025-12-06 06:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:52:30.705461291 +0000 UTC m=+4390.355728943" watchObservedRunningTime="2025-12-06 06:52:30.712020469 +0000 UTC m=+4390.362288101" Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.729063 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-jpcvt"] Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.731600 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.837069 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42cp7\" (UniqueName: \"kubernetes.io/projected/3348cd26-fbd1-4f78-bbbe-1c18203b712e-kube-api-access-42cp7\") pod \"crc-debug-jpcvt\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.837579 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3348cd26-fbd1-4f78-bbbe-1c18203b712e-host\") pod \"crc-debug-jpcvt\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.939535 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42cp7\" (UniqueName: \"kubernetes.io/projected/3348cd26-fbd1-4f78-bbbe-1c18203b712e-kube-api-access-42cp7\") pod \"crc-debug-jpcvt\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.940215 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3348cd26-fbd1-4f78-bbbe-1c18203b712e-host\") pod \"crc-debug-jpcvt\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.940360 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3348cd26-fbd1-4f78-bbbe-1c18203b712e-host\") pod \"crc-debug-jpcvt\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:33 crc kubenswrapper[4957]: I1206 06:52:33.966093 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42cp7\" (UniqueName: \"kubernetes.io/projected/3348cd26-fbd1-4f78-bbbe-1c18203b712e-kube-api-access-42cp7\") pod \"crc-debug-jpcvt\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:34 crc kubenswrapper[4957]: I1206 06:52:34.049032 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:52:34 crc kubenswrapper[4957]: W1206 06:52:34.076561 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3348cd26_fbd1_4f78_bbbe_1c18203b712e.slice/crio-c3c736246f73b3974368a744b1ac371994fd1c16299276b1c77de0debee9d6d6 WatchSource:0}: Error finding container c3c736246f73b3974368a744b1ac371994fd1c16299276b1c77de0debee9d6d6: Status 404 returned error can't find the container with id c3c736246f73b3974368a744b1ac371994fd1c16299276b1c77de0debee9d6d6 Dec 06 06:52:34 crc kubenswrapper[4957]: I1206 06:52:34.733254 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" event={"ID":"3348cd26-fbd1-4f78-bbbe-1c18203b712e","Type":"ContainerStarted","Data":"877ee65d77fb25c6334bc121733e7b3f5ac3aedd628c5bf1d8dcf81f6f40512b"} Dec 06 06:52:34 crc kubenswrapper[4957]: I1206 06:52:34.733948 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" event={"ID":"3348cd26-fbd1-4f78-bbbe-1c18203b712e","Type":"ContainerStarted","Data":"c3c736246f73b3974368a744b1ac371994fd1c16299276b1c77de0debee9d6d6"} Dec 06 06:52:34 crc kubenswrapper[4957]: I1206 06:52:34.755789 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" podStartSLOduration=1.755763255 podStartE2EDuration="1.755763255s" podCreationTimestamp="2025-12-06 06:52:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 06:52:34.751379826 +0000 UTC m=+4394.401647468" watchObservedRunningTime="2025-12-06 06:52:34.755763255 +0000 UTC m=+4394.406030897" Dec 06 06:52:40 crc kubenswrapper[4957]: I1206 06:52:40.669156 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:52:40 crc kubenswrapper[4957]: E1206 06:52:40.669899 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.102559 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-24226"] Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.104735 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.119561 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24226"] Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.187282 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6rrm\" (UniqueName: \"kubernetes.io/projected/202a542f-4f0f-4aa5-88f4-176e1d651823-kube-api-access-w6rrm\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.187343 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-utilities\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.187742 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-catalog-content\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.289129 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-catalog-content\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.289214 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6rrm\" (UniqueName: \"kubernetes.io/projected/202a542f-4f0f-4aa5-88f4-176e1d651823-kube-api-access-w6rrm\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.289236 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-utilities\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.289749 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-catalog-content\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.289763 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-utilities\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:41 crc kubenswrapper[4957]: I1206 06:52:41.314446 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6rrm\" (UniqueName: \"kubernetes.io/projected/202a542f-4f0f-4aa5-88f4-176e1d651823-kube-api-access-w6rrm\") pod \"certified-operators-24226\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:42 crc kubenswrapper[4957]: I1206 06:52:42.622498 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:43 crc kubenswrapper[4957]: I1206 06:52:43.178864 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24226"] Dec 06 06:52:43 crc kubenswrapper[4957]: I1206 06:52:43.811318 4957 generic.go:334] "Generic (PLEG): container finished" podID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerID="d6f30782973adb42c1f3c7e5db505f9adc0bddce17009156f5cb29ea7705aaba" exitCode=0 Dec 06 06:52:43 crc kubenswrapper[4957]: I1206 06:52:43.811385 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24226" event={"ID":"202a542f-4f0f-4aa5-88f4-176e1d651823","Type":"ContainerDied","Data":"d6f30782973adb42c1f3c7e5db505f9adc0bddce17009156f5cb29ea7705aaba"} Dec 06 06:52:43 crc kubenswrapper[4957]: I1206 06:52:43.811450 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24226" event={"ID":"202a542f-4f0f-4aa5-88f4-176e1d651823","Type":"ContainerStarted","Data":"b85268e81bd48eb8d8f310bea50958c3b2399b4696efca6905ccd2f37d8779bd"} Dec 06 06:52:43 crc kubenswrapper[4957]: I1206 06:52:43.813689 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:52:44 crc kubenswrapper[4957]: I1206 06:52:44.822640 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24226" event={"ID":"202a542f-4f0f-4aa5-88f4-176e1d651823","Type":"ContainerStarted","Data":"988be5e12e3e7401aeab006c6ef22328a85eb2fd6218737e738acef23f34d5ab"} Dec 06 06:52:45 crc kubenswrapper[4957]: I1206 06:52:45.837250 4957 generic.go:334] "Generic (PLEG): container finished" podID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerID="988be5e12e3e7401aeab006c6ef22328a85eb2fd6218737e738acef23f34d5ab" exitCode=0 Dec 06 06:52:45 crc kubenswrapper[4957]: I1206 06:52:45.837562 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24226" event={"ID":"202a542f-4f0f-4aa5-88f4-176e1d651823","Type":"ContainerDied","Data":"988be5e12e3e7401aeab006c6ef22328a85eb2fd6218737e738acef23f34d5ab"} Dec 06 06:52:47 crc kubenswrapper[4957]: I1206 06:52:47.872492 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24226" event={"ID":"202a542f-4f0f-4aa5-88f4-176e1d651823","Type":"ContainerStarted","Data":"9780a8a1e934f65140da381dc78bae3297b323cd113453e07142788bd2ab1c97"} Dec 06 06:52:47 crc kubenswrapper[4957]: I1206 06:52:47.892001 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-24226" podStartSLOduration=4.43503813 podStartE2EDuration="6.8919813s" podCreationTimestamp="2025-12-06 06:52:41 +0000 UTC" firstStartedPulling="2025-12-06 06:52:43.813465237 +0000 UTC m=+4403.463732869" lastFinishedPulling="2025-12-06 06:52:46.270408407 +0000 UTC m=+4405.920676039" observedRunningTime="2025-12-06 06:52:47.888947298 +0000 UTC m=+4407.539214960" watchObservedRunningTime="2025-12-06 06:52:47.8919813 +0000 UTC m=+4407.542248932" Dec 06 06:52:52 crc kubenswrapper[4957]: I1206 06:52:52.623924 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:52 crc kubenswrapper[4957]: I1206 06:52:52.625676 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:52 crc kubenswrapper[4957]: I1206 06:52:52.662873 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:52:52 crc kubenswrapper[4957]: E1206 06:52:52.663410 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:52:52 crc kubenswrapper[4957]: I1206 06:52:52.679115 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:52 crc kubenswrapper[4957]: I1206 06:52:52.962574 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:53 crc kubenswrapper[4957]: I1206 06:52:53.013422 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24226"] Dec 06 06:52:54 crc kubenswrapper[4957]: I1206 06:52:54.935447 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-24226" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="registry-server" containerID="cri-o://9780a8a1e934f65140da381dc78bae3297b323cd113453e07142788bd2ab1c97" gracePeriod=2 Dec 06 06:52:55 crc kubenswrapper[4957]: I1206 06:52:55.973293 4957 generic.go:334] "Generic (PLEG): container finished" podID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerID="9780a8a1e934f65140da381dc78bae3297b323cd113453e07142788bd2ab1c97" exitCode=0 Dec 06 06:52:55 crc kubenswrapper[4957]: I1206 06:52:55.973340 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24226" event={"ID":"202a542f-4f0f-4aa5-88f4-176e1d651823","Type":"ContainerDied","Data":"9780a8a1e934f65140da381dc78bae3297b323cd113453e07142788bd2ab1c97"} Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.615104 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.688719 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-catalog-content\") pod \"202a542f-4f0f-4aa5-88f4-176e1d651823\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.688806 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-utilities\") pod \"202a542f-4f0f-4aa5-88f4-176e1d651823\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.689039 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6rrm\" (UniqueName: \"kubernetes.io/projected/202a542f-4f0f-4aa5-88f4-176e1d651823-kube-api-access-w6rrm\") pod \"202a542f-4f0f-4aa5-88f4-176e1d651823\" (UID: \"202a542f-4f0f-4aa5-88f4-176e1d651823\") " Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.692342 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-utilities" (OuterVolumeSpecName: "utilities") pod "202a542f-4f0f-4aa5-88f4-176e1d651823" (UID: "202a542f-4f0f-4aa5-88f4-176e1d651823"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.699472 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/202a542f-4f0f-4aa5-88f4-176e1d651823-kube-api-access-w6rrm" (OuterVolumeSpecName: "kube-api-access-w6rrm") pod "202a542f-4f0f-4aa5-88f4-176e1d651823" (UID: "202a542f-4f0f-4aa5-88f4-176e1d651823"). InnerVolumeSpecName "kube-api-access-w6rrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.758675 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "202a542f-4f0f-4aa5-88f4-176e1d651823" (UID: "202a542f-4f0f-4aa5-88f4-176e1d651823"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.791737 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6rrm\" (UniqueName: \"kubernetes.io/projected/202a542f-4f0f-4aa5-88f4-176e1d651823-kube-api-access-w6rrm\") on node \"crc\" DevicePath \"\"" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.791775 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.791785 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202a542f-4f0f-4aa5-88f4-176e1d651823-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.983346 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24226" event={"ID":"202a542f-4f0f-4aa5-88f4-176e1d651823","Type":"ContainerDied","Data":"b85268e81bd48eb8d8f310bea50958c3b2399b4696efca6905ccd2f37d8779bd"} Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.983404 4957 scope.go:117] "RemoveContainer" containerID="9780a8a1e934f65140da381dc78bae3297b323cd113453e07142788bd2ab1c97" Dec 06 06:52:56 crc kubenswrapper[4957]: I1206 06:52:56.983442 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24226" Dec 06 06:52:57 crc kubenswrapper[4957]: I1206 06:52:57.021378 4957 scope.go:117] "RemoveContainer" containerID="988be5e12e3e7401aeab006c6ef22328a85eb2fd6218737e738acef23f34d5ab" Dec 06 06:52:57 crc kubenswrapper[4957]: I1206 06:52:57.034207 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24226"] Dec 06 06:52:57 crc kubenswrapper[4957]: I1206 06:52:57.047545 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-24226"] Dec 06 06:52:57 crc kubenswrapper[4957]: I1206 06:52:57.054728 4957 scope.go:117] "RemoveContainer" containerID="d6f30782973adb42c1f3c7e5db505f9adc0bddce17009156f5cb29ea7705aaba" Dec 06 06:52:58 crc kubenswrapper[4957]: I1206 06:52:58.680110 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" path="/var/lib/kubelet/pods/202a542f-4f0f-4aa5-88f4-176e1d651823/volumes" Dec 06 06:53:03 crc kubenswrapper[4957]: I1206 06:53:03.662208 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:53:03 crc kubenswrapper[4957]: E1206 06:53:03.663014 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:53:07 crc kubenswrapper[4957]: I1206 06:53:07.086223 4957 generic.go:334] "Generic (PLEG): container finished" podID="3348cd26-fbd1-4f78-bbbe-1c18203b712e" containerID="877ee65d77fb25c6334bc121733e7b3f5ac3aedd628c5bf1d8dcf81f6f40512b" exitCode=0 Dec 06 06:53:07 crc kubenswrapper[4957]: I1206 06:53:07.086334 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" event={"ID":"3348cd26-fbd1-4f78-bbbe-1c18203b712e","Type":"ContainerDied","Data":"877ee65d77fb25c6334bc121733e7b3f5ac3aedd628c5bf1d8dcf81f6f40512b"} Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.480798 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.515136 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-jpcvt"] Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.523630 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-jpcvt"] Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.607986 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3348cd26-fbd1-4f78-bbbe-1c18203b712e-host\") pod \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.608053 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42cp7\" (UniqueName: \"kubernetes.io/projected/3348cd26-fbd1-4f78-bbbe-1c18203b712e-kube-api-access-42cp7\") pod \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\" (UID: \"3348cd26-fbd1-4f78-bbbe-1c18203b712e\") " Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.608267 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3348cd26-fbd1-4f78-bbbe-1c18203b712e-host" (OuterVolumeSpecName: "host") pod "3348cd26-fbd1-4f78-bbbe-1c18203b712e" (UID: "3348cd26-fbd1-4f78-bbbe-1c18203b712e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.608494 4957 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3348cd26-fbd1-4f78-bbbe-1c18203b712e-host\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.615119 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3348cd26-fbd1-4f78-bbbe-1c18203b712e-kube-api-access-42cp7" (OuterVolumeSpecName: "kube-api-access-42cp7") pod "3348cd26-fbd1-4f78-bbbe-1c18203b712e" (UID: "3348cd26-fbd1-4f78-bbbe-1c18203b712e"). InnerVolumeSpecName "kube-api-access-42cp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.675541 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3348cd26-fbd1-4f78-bbbe-1c18203b712e" path="/var/lib/kubelet/pods/3348cd26-fbd1-4f78-bbbe-1c18203b712e/volumes" Dec 06 06:53:08 crc kubenswrapper[4957]: I1206 06:53:08.710482 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42cp7\" (UniqueName: \"kubernetes.io/projected/3348cd26-fbd1-4f78-bbbe-1c18203b712e-kube-api-access-42cp7\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.107367 4957 scope.go:117] "RemoveContainer" containerID="877ee65d77fb25c6334bc121733e7b3f5ac3aedd628c5bf1d8dcf81f6f40512b" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.107503 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-jpcvt" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.720896 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-54j8x"] Dec 06 06:53:09 crc kubenswrapper[4957]: E1206 06:53:09.721767 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="extract-utilities" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.721945 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="extract-utilities" Dec 06 06:53:09 crc kubenswrapper[4957]: E1206 06:53:09.722025 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="extract-content" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.722106 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="extract-content" Dec 06 06:53:09 crc kubenswrapper[4957]: E1206 06:53:09.722165 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="registry-server" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.722229 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="registry-server" Dec 06 06:53:09 crc kubenswrapper[4957]: E1206 06:53:09.722305 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3348cd26-fbd1-4f78-bbbe-1c18203b712e" containerName="container-00" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.722359 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3348cd26-fbd1-4f78-bbbe-1c18203b712e" containerName="container-00" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.722592 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="202a542f-4f0f-4aa5-88f4-176e1d651823" containerName="registry-server" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.722658 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3348cd26-fbd1-4f78-bbbe-1c18203b712e" containerName="container-00" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.723340 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.831406 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2jhm\" (UniqueName: \"kubernetes.io/projected/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-kube-api-access-k2jhm\") pod \"crc-debug-54j8x\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.831676 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-host\") pod \"crc-debug-54j8x\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.933368 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-host\") pod \"crc-debug-54j8x\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.933484 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2jhm\" (UniqueName: \"kubernetes.io/projected/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-kube-api-access-k2jhm\") pod \"crc-debug-54j8x\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.933541 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-host\") pod \"crc-debug-54j8x\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:09 crc kubenswrapper[4957]: I1206 06:53:09.950062 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2jhm\" (UniqueName: \"kubernetes.io/projected/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-kube-api-access-k2jhm\") pod \"crc-debug-54j8x\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:10 crc kubenswrapper[4957]: I1206 06:53:10.039176 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:10 crc kubenswrapper[4957]: I1206 06:53:10.121689 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/crc-debug-54j8x" event={"ID":"3e03e4b5-6f75-4477-ad84-a8662fbb40e5","Type":"ContainerStarted","Data":"9411244ee8c1f4307c6cf8b296de0c57aef530c81098ef91666b2c2ffb8fbc75"} Dec 06 06:53:11 crc kubenswrapper[4957]: I1206 06:53:11.139708 4957 generic.go:334] "Generic (PLEG): container finished" podID="3e03e4b5-6f75-4477-ad84-a8662fbb40e5" containerID="18c1afe198f9a79139c77fcfd4d3ecf7a2f150a014ab75e8e60e6a12f761f305" exitCode=0 Dec 06 06:53:11 crc kubenswrapper[4957]: I1206 06:53:11.139762 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/crc-debug-54j8x" event={"ID":"3e03e4b5-6f75-4477-ad84-a8662fbb40e5","Type":"ContainerDied","Data":"18c1afe198f9a79139c77fcfd4d3ecf7a2f150a014ab75e8e60e6a12f761f305"} Dec 06 06:53:11 crc kubenswrapper[4957]: I1206 06:53:11.623722 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-54j8x"] Dec 06 06:53:11 crc kubenswrapper[4957]: I1206 06:53:11.631218 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-54j8x"] Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.275708 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.391821 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-host\") pod \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.392062 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2jhm\" (UniqueName: \"kubernetes.io/projected/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-kube-api-access-k2jhm\") pod \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\" (UID: \"3e03e4b5-6f75-4477-ad84-a8662fbb40e5\") " Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.392218 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-host" (OuterVolumeSpecName: "host") pod "3e03e4b5-6f75-4477-ad84-a8662fbb40e5" (UID: "3e03e4b5-6f75-4477-ad84-a8662fbb40e5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.392570 4957 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-host\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.400243 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-kube-api-access-k2jhm" (OuterVolumeSpecName: "kube-api-access-k2jhm") pod "3e03e4b5-6f75-4477-ad84-a8662fbb40e5" (UID: "3e03e4b5-6f75-4477-ad84-a8662fbb40e5"). InnerVolumeSpecName "kube-api-access-k2jhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.494623 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2jhm\" (UniqueName: \"kubernetes.io/projected/3e03e4b5-6f75-4477-ad84-a8662fbb40e5-kube-api-access-k2jhm\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.673394 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e03e4b5-6f75-4477-ad84-a8662fbb40e5" path="/var/lib/kubelet/pods/3e03e4b5-6f75-4477-ad84-a8662fbb40e5/volumes" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.807307 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-rcz7b"] Dec 06 06:53:12 crc kubenswrapper[4957]: E1206 06:53:12.808346 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e03e4b5-6f75-4477-ad84-a8662fbb40e5" containerName="container-00" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.808394 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e03e4b5-6f75-4477-ad84-a8662fbb40e5" containerName="container-00" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.808700 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e03e4b5-6f75-4477-ad84-a8662fbb40e5" containerName="container-00" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.809450 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.901586 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9d3398d-db92-4c25-b214-c3eff574a032-host\") pod \"crc-debug-rcz7b\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:12 crc kubenswrapper[4957]: I1206 06:53:12.901625 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z68w\" (UniqueName: \"kubernetes.io/projected/d9d3398d-db92-4c25-b214-c3eff574a032-kube-api-access-9z68w\") pod \"crc-debug-rcz7b\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:13 crc kubenswrapper[4957]: I1206 06:53:13.004052 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9d3398d-db92-4c25-b214-c3eff574a032-host\") pod \"crc-debug-rcz7b\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:13 crc kubenswrapper[4957]: I1206 06:53:13.004115 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z68w\" (UniqueName: \"kubernetes.io/projected/d9d3398d-db92-4c25-b214-c3eff574a032-kube-api-access-9z68w\") pod \"crc-debug-rcz7b\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:13 crc kubenswrapper[4957]: I1206 06:53:13.004311 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9d3398d-db92-4c25-b214-c3eff574a032-host\") pod \"crc-debug-rcz7b\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:13 crc kubenswrapper[4957]: I1206 06:53:13.024292 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z68w\" (UniqueName: \"kubernetes.io/projected/d9d3398d-db92-4c25-b214-c3eff574a032-kube-api-access-9z68w\") pod \"crc-debug-rcz7b\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:13 crc kubenswrapper[4957]: I1206 06:53:13.129823 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:13 crc kubenswrapper[4957]: W1206 06:53:13.159083 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9d3398d_db92_4c25_b214_c3eff574a032.slice/crio-8ef65665901f8b5e567de1b031e3c7e1384f5d6b34c2e2de0b37584cd1fb0b3d WatchSource:0}: Error finding container 8ef65665901f8b5e567de1b031e3c7e1384f5d6b34c2e2de0b37584cd1fb0b3d: Status 404 returned error can't find the container with id 8ef65665901f8b5e567de1b031e3c7e1384f5d6b34c2e2de0b37584cd1fb0b3d Dec 06 06:53:13 crc kubenswrapper[4957]: I1206 06:53:13.159995 4957 scope.go:117] "RemoveContainer" containerID="18c1afe198f9a79139c77fcfd4d3ecf7a2f150a014ab75e8e60e6a12f761f305" Dec 06 06:53:13 crc kubenswrapper[4957]: I1206 06:53:13.160024 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-54j8x" Dec 06 06:53:14 crc kubenswrapper[4957]: I1206 06:53:14.174424 4957 generic.go:334] "Generic (PLEG): container finished" podID="d9d3398d-db92-4c25-b214-c3eff574a032" containerID="a95452857c0513aeb0f8d02553e84ce5fc71c97cd891e90cb49f203fdda587c4" exitCode=0 Dec 06 06:53:14 crc kubenswrapper[4957]: I1206 06:53:14.174563 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" event={"ID":"d9d3398d-db92-4c25-b214-c3eff574a032","Type":"ContainerDied","Data":"a95452857c0513aeb0f8d02553e84ce5fc71c97cd891e90cb49f203fdda587c4"} Dec 06 06:53:14 crc kubenswrapper[4957]: I1206 06:53:14.175127 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" event={"ID":"d9d3398d-db92-4c25-b214-c3eff574a032","Type":"ContainerStarted","Data":"8ef65665901f8b5e567de1b031e3c7e1384f5d6b34c2e2de0b37584cd1fb0b3d"} Dec 06 06:53:14 crc kubenswrapper[4957]: I1206 06:53:14.228389 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-rcz7b"] Dec 06 06:53:14 crc kubenswrapper[4957]: I1206 06:53:14.238817 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9qkxf/crc-debug-rcz7b"] Dec 06 06:53:15 crc kubenswrapper[4957]: I1206 06:53:15.294329 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:15 crc kubenswrapper[4957]: I1206 06:53:15.461297 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z68w\" (UniqueName: \"kubernetes.io/projected/d9d3398d-db92-4c25-b214-c3eff574a032-kube-api-access-9z68w\") pod \"d9d3398d-db92-4c25-b214-c3eff574a032\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " Dec 06 06:53:15 crc kubenswrapper[4957]: I1206 06:53:15.461620 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9d3398d-db92-4c25-b214-c3eff574a032-host\") pod \"d9d3398d-db92-4c25-b214-c3eff574a032\" (UID: \"d9d3398d-db92-4c25-b214-c3eff574a032\") " Dec 06 06:53:15 crc kubenswrapper[4957]: I1206 06:53:15.461889 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9d3398d-db92-4c25-b214-c3eff574a032-host" (OuterVolumeSpecName: "host") pod "d9d3398d-db92-4c25-b214-c3eff574a032" (UID: "d9d3398d-db92-4c25-b214-c3eff574a032"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 06:53:15 crc kubenswrapper[4957]: I1206 06:53:15.462613 4957 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9d3398d-db92-4c25-b214-c3eff574a032-host\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:15 crc kubenswrapper[4957]: I1206 06:53:15.469822 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d3398d-db92-4c25-b214-c3eff574a032-kube-api-access-9z68w" (OuterVolumeSpecName: "kube-api-access-9z68w") pod "d9d3398d-db92-4c25-b214-c3eff574a032" (UID: "d9d3398d-db92-4c25-b214-c3eff574a032"). InnerVolumeSpecName "kube-api-access-9z68w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:53:15 crc kubenswrapper[4957]: I1206 06:53:15.564643 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z68w\" (UniqueName: \"kubernetes.io/projected/d9d3398d-db92-4c25-b214-c3eff574a032-kube-api-access-9z68w\") on node \"crc\" DevicePath \"\"" Dec 06 06:53:16 crc kubenswrapper[4957]: I1206 06:53:16.198016 4957 scope.go:117] "RemoveContainer" containerID="a95452857c0513aeb0f8d02553e84ce5fc71c97cd891e90cb49f203fdda587c4" Dec 06 06:53:16 crc kubenswrapper[4957]: I1206 06:53:16.198098 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/crc-debug-rcz7b" Dec 06 06:53:16 crc kubenswrapper[4957]: I1206 06:53:16.673488 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d3398d-db92-4c25-b214-c3eff574a032" path="/var/lib/kubelet/pods/d9d3398d-db92-4c25-b214-c3eff574a032/volumes" Dec 06 06:53:18 crc kubenswrapper[4957]: I1206 06:53:18.662470 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:53:18 crc kubenswrapper[4957]: E1206 06:53:18.663186 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:53:32 crc kubenswrapper[4957]: I1206 06:53:32.662169 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:53:32 crc kubenswrapper[4957]: E1206 06:53:32.662947 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:53:40 crc kubenswrapper[4957]: I1206 06:53:40.957416 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c9f9fb958-p7bk5_59b9c817-9c5c-4f08-9ca9-9d54e1d0912b/barbican-api/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.041134 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c9f9fb958-p7bk5_59b9c817-9c5c-4f08-9ca9-9d54e1d0912b/barbican-api-log/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.214982 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bb9fc9b94-kknvp_e3907b98-0bd2-4ad9-bb72-d1f651c16f68/barbican-keystone-listener-log/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.238125 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bb9fc9b94-kknvp_e3907b98-0bd2-4ad9-bb72-d1f651c16f68/barbican-keystone-listener/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.473933 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5684df76ff-6tw89_b1c9d4c1-173d-4db7-9986-2873c3280e2a/barbican-worker-log/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.486323 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5684df76ff-6tw89_b1c9d4c1-173d-4db7-9986-2873c3280e2a/barbican-worker/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.557872 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gr5v9_46b05e85-6b4c-4aae-ab6b-39e026f48d49/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.675264 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/ceilometer-central-agent/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.847204 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/ceilometer-notification-agent/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.881978 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/proxy-httpd/0.log" Dec 06 06:53:41 crc kubenswrapper[4957]: I1206 06:53:41.933171 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_963f3a5a-72a1-4fed-9ca8-edc041e612c3/sg-core/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.102583 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c21557b2-c3ad-497a-ad42-ca81fff434bf/cinder-api/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.177310 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c21557b2-c3ad-497a-ad42-ca81fff434bf/cinder-api-log/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.273100 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_78ea9b70-3dde-4183-9ce4-c61326745bf2/cinder-scheduler/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.381400 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xrvz5_a432bf6f-c43f-4b87-af4b-2af6281f1c25/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.393294 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_78ea9b70-3dde-4183-9ce4-c61326745bf2/probe/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.573889 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9wdwc_2cb8adc2-7a82-4969-93d3-abac49e01d24/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.607927 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-7vzqk_3bac0712-aa19-4f2f-9645-eecdf3d19c5f/init/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.822670 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-7vzqk_3bac0712-aa19-4f2f-9645-eecdf3d19c5f/init/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.912088 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-7vzqk_3bac0712-aa19-4f2f-9645-eecdf3d19c5f/dnsmasq-dns/0.log" Dec 06 06:53:42 crc kubenswrapper[4957]: I1206 06:53:42.928707 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-shvb7_2dc93721-02c2-4553-960f-71955c73b548/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:43 crc kubenswrapper[4957]: I1206 06:53:43.310060 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2319a013-4227-4643-9b7f-d60b4d501af7/glance-log/0.log" Dec 06 06:53:43 crc kubenswrapper[4957]: I1206 06:53:43.382128 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2319a013-4227-4643-9b7f-d60b4d501af7/glance-httpd/0.log" Dec 06 06:53:43 crc kubenswrapper[4957]: I1206 06:53:43.475283 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e8deee48-5aef-4889-b7ab-b7f23144cbc8/glance-log/0.log" Dec 06 06:53:43 crc kubenswrapper[4957]: I1206 06:53:43.494880 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e8deee48-5aef-4889-b7ab-b7f23144cbc8/glance-httpd/0.log" Dec 06 06:53:43 crc kubenswrapper[4957]: I1206 06:53:43.758895 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c76d956-bsr6l_cb06a82d-1610-4985-8168-c95f32ad2ff6/horizon/0.log" Dec 06 06:53:43 crc kubenswrapper[4957]: I1206 06:53:43.884423 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t9hhz_26574c15-db39-470e-b739-cf03859b46aa/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:43 crc kubenswrapper[4957]: I1206 06:53:43.997484 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-2n487_87515d38-f8c8-4dd5-9373-fd2bd9ef7992/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.025993 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c76d956-bsr6l_cb06a82d-1610-4985-8168-c95f32ad2ff6/horizon-log/0.log" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.192937 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6a476151-cb3f-4502-8b9a-be049c359804/kube-state-metrics/0.log" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.198823 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7c7fdbc8b-njnq6_95511bb1-a0e3-4761-aba1-fd10b37df59f/keystone-api/0.log" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.362142 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-jzkqz_5df54304-6d88-4027-ae21-40aaa328470f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.647988 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-575d75c97c-wblw4_a7c5d2f8-fce6-4689-8090-9ccbc8bb795e/neutron-api/0.log" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.653517 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-575d75c97c-wblw4_a7c5d2f8-fce6-4689-8090-9ccbc8bb795e/neutron-httpd/0.log" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.662557 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:53:44 crc kubenswrapper[4957]: E1206 06:53:44.662758 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:53:44 crc kubenswrapper[4957]: I1206 06:53:44.723460 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-cqm4n_a0ec714a-36ad-46f4-aeca-a912266812f2/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:45 crc kubenswrapper[4957]: I1206 06:53:45.329281 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1ae0f500-0ce2-4374-afe8-cc65e465ac26/nova-api-log/0.log" Dec 06 06:53:45 crc kubenswrapper[4957]: I1206 06:53:45.439192 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_9ce84760-5e40-4ea1-8bb2-81dc593c757a/nova-cell0-conductor-conductor/0.log" Dec 06 06:53:45 crc kubenswrapper[4957]: I1206 06:53:45.836127 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_1ae0f500-0ce2-4374-afe8-cc65e465ac26/nova-api-api/0.log" Dec 06 06:53:45 crc kubenswrapper[4957]: I1206 06:53:45.856693 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_6d950863-c9e9-4bc9-ae48-c8d1e63ad75c/nova-cell1-conductor-conductor/0.log" Dec 06 06:53:45 crc kubenswrapper[4957]: I1206 06:53:45.877724 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ecb88cbc-d465-49e9-9aaf-31aa46e18ae2/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 06:53:46 crc kubenswrapper[4957]: I1206 06:53:46.151237 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8dt69_312ec0f6-c580-4326-ba1a-b80706d87299/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:46 crc kubenswrapper[4957]: I1206 06:53:46.242287 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_67d6edc5-c0f1-4466-8b5b-99cd3bc06d17/nova-metadata-log/0.log" Dec 06 06:53:46 crc kubenswrapper[4957]: I1206 06:53:46.591529 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c10af3da-4556-4f2f-bfb8-886a6a93bfc4/mysql-bootstrap/0.log" Dec 06 06:53:46 crc kubenswrapper[4957]: I1206 06:53:46.643729 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c683a5a8-170a-4880-b9fa-7b9635ce0751/nova-scheduler-scheduler/0.log" Dec 06 06:53:46 crc kubenswrapper[4957]: I1206 06:53:46.785958 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c10af3da-4556-4f2f-bfb8-886a6a93bfc4/mysql-bootstrap/0.log" Dec 06 06:53:46 crc kubenswrapper[4957]: I1206 06:53:46.820057 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c10af3da-4556-4f2f-bfb8-886a6a93bfc4/galera/0.log" Dec 06 06:53:46 crc kubenswrapper[4957]: I1206 06:53:46.993673 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d87bf6f-a3d1-41ce-ba76-ed55aec8645e/mysql-bootstrap/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.206491 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d87bf6f-a3d1-41ce-ba76-ed55aec8645e/mysql-bootstrap/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.237171 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3d87bf6f-a3d1-41ce-ba76-ed55aec8645e/galera/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.429597 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_663bc0fb-e06b-4c91-8a36-a97d32fa63f3/openstackclient/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.438063 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-q7r5j_e5adfb1c-2a78-42af-af7e-419bf4d73bd4/openstack-network-exporter/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.648502 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovsdb-server-init/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.783727 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_67d6edc5-c0f1-4466-8b5b-99cd3bc06d17/nova-metadata-metadata/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.834106 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovsdb-server-init/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.837953 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovs-vswitchd/0.log" Dec 06 06:53:47 crc kubenswrapper[4957]: I1206 06:53:47.924713 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qhgsm_0dfa83fc-3365-4d78-980f-23a2c71711f7/ovsdb-server/0.log" Dec 06 06:53:48 crc kubenswrapper[4957]: I1206 06:53:48.683369 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pmd7t_bf193218-2a15-4cf5-8ba0-59dc939445ff/ovn-controller/0.log" Dec 06 06:53:48 crc kubenswrapper[4957]: I1206 06:53:48.757134 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8bwfh_1eaa20e5-44dd-418a-94c2-fac0a16ef54a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:48 crc kubenswrapper[4957]: I1206 06:53:48.938106 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7262d5d8-b26d-4871-a3c7-9ff1f2c4009f/openstack-network-exporter/0.log" Dec 06 06:53:48 crc kubenswrapper[4957]: I1206 06:53:48.943427 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7262d5d8-b26d-4871-a3c7-9ff1f2c4009f/ovn-northd/0.log" Dec 06 06:53:49 crc kubenswrapper[4957]: I1206 06:53:49.146142 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d6ec6d60-cadc-488c-bb21-514a7acbfc70/openstack-network-exporter/0.log" Dec 06 06:53:49 crc kubenswrapper[4957]: I1206 06:53:49.246137 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6c710a2d-a836-4154-9b55-70e3d6ad61a1/openstack-network-exporter/0.log" Dec 06 06:53:49 crc kubenswrapper[4957]: I1206 06:53:49.253842 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d6ec6d60-cadc-488c-bb21-514a7acbfc70/ovsdbserver-nb/0.log" Dec 06 06:53:49 crc kubenswrapper[4957]: I1206 06:53:49.367068 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6c710a2d-a836-4154-9b55-70e3d6ad61a1/ovsdbserver-sb/0.log" Dec 06 06:53:49 crc kubenswrapper[4957]: I1206 06:53:49.536075 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69dd5567fb-lm2jl_8cada75b-b0d4-41c2-8d1f-3f146e28651d/placement-api/0.log" Dec 06 06:53:49 crc kubenswrapper[4957]: I1206 06:53:49.602405 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69dd5567fb-lm2jl_8cada75b-b0d4-41c2-8d1f-3f146e28651d/placement-log/0.log" Dec 06 06:53:49 crc kubenswrapper[4957]: I1206 06:53:49.654607 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a28833e8-d161-4dd0-816e-b476ab8d473b/setup-container/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.280463 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a28833e8-d161-4dd0-816e-b476ab8d473b/setup-container/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.328174 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8504457c-7804-41fc-ac00-c2583a75137b/setup-container/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.328576 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a28833e8-d161-4dd0-816e-b476ab8d473b/rabbitmq/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.500825 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8504457c-7804-41fc-ac00-c2583a75137b/rabbitmq/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.540523 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8504457c-7804-41fc-ac00-c2583a75137b/setup-container/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.586942 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-bq4wv_101a78c1-076b-41be-a2bf-04f2f987924b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.746181 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-rc87w_6206b723-2b44-49ab-a3ec-93ce9d406ae8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.816437 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jt982_a75e174a-7ba2-44d3-9d5d-060259173a48/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:50 crc kubenswrapper[4957]: I1206 06:53:50.962171 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-rk8hd_9ca55297-0691-458e-85c2-a408a031c485/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.125229 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-c7hml_33b5c66b-1a29-44b1-9863-08dc69965dac/ssh-known-hosts-edpm-deployment/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.361062 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5789568449-4t258_475d23a9-0dda-4ba5-8af9-108574997a4f/proxy-server/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.398260 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wm72g_43f9d83d-effb-4bc6-8550-6e4c32adaf8d/swift-ring-rebalance/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.459163 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5789568449-4t258_475d23a9-0dda-4ba5-8af9-108574997a4f/proxy-httpd/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.619285 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-auditor/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.646808 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-reaper/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.760386 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-replicator/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.845773 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-auditor/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.856437 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/account-server/0.log" Dec 06 06:53:51 crc kubenswrapper[4957]: I1206 06:53:51.954871 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-replicator/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.002605 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-server/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.026202 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/container-updater/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.122182 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-auditor/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.181067 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-expirer/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.217717 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-replicator/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.243478 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-server/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.338029 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/object-updater/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.394593 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/rsync/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.450909 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_72358664-e14f-482d-bd35-c807922b525f/swift-recon-cron/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.656732 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_469a494f-1ceb-4d9f-bfe5-e0856e88f5bd/tempest-tests-tempest-tests-runner/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.662749 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g78cd_37341858-eaac-4c32-a5cc-7b06843f4fc2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.840535 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d42a9079-5c1e-424a-b79b-d469afb272c3/test-operator-logs-container/0.log" Dec 06 06:53:52 crc kubenswrapper[4957]: I1206 06:53:52.919496 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-f799z_7f74c7c3-a665-4958-aaad-15ed4e962fd0/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 06:53:59 crc kubenswrapper[4957]: I1206 06:53:59.662116 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:53:59 crc kubenswrapper[4957]: E1206 06:53:59.663033 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:54:04 crc kubenswrapper[4957]: I1206 06:54:04.016084 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_528a2b84-7047-444d-997e-4a92e0b5dbf9/memcached/0.log" Dec 06 06:54:12 crc kubenswrapper[4957]: I1206 06:54:12.661848 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:54:12 crc kubenswrapper[4957]: E1206 06:54:12.662553 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:54:19 crc kubenswrapper[4957]: I1206 06:54:19.674594 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rpf52_6276ae06-7d72-409a-81f2-70ba70bf3d86/kube-rbac-proxy/0.log" Dec 06 06:54:19 crc kubenswrapper[4957]: I1206 06:54:19.780600 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rpf52_6276ae06-7d72-409a-81f2-70ba70bf3d86/manager/0.log" Dec 06 06:54:19 crc kubenswrapper[4957]: I1206 06:54:19.909910 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mzk64_bf2466dc-9cd3-4be7-9d17-846eccda13a8/kube-rbac-proxy/0.log" Dec 06 06:54:19 crc kubenswrapper[4957]: I1206 06:54:19.977915 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mzk64_bf2466dc-9cd3-4be7-9d17-846eccda13a8/manager/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.090996 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-9m6rj_e7ea0507-7466-4c88-a20b-2c936c6cdccb/kube-rbac-proxy/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.113320 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-9m6rj_e7ea0507-7466-4c88-a20b-2c936c6cdccb/manager/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.177226 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/util/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.342218 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/pull/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.346518 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/pull/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.355159 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/util/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.539742 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/extract/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.549592 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/util/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.584242 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8w9v45_9b230128-8dcd-4f69-9bcc-e57a23bd3ef6/pull/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.736500 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d7r5m_c5d6f879-3842-4481-88bd-bb090eecd9fc/kube-rbac-proxy/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.805958 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d7r5m_c5d6f879-3842-4481-88bd-bb090eecd9fc/manager/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.896690 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rs7b8_ce46ec44-fc83-46e0-9f85-159e05452f0f/kube-rbac-proxy/0.log" Dec 06 06:54:20 crc kubenswrapper[4957]: I1206 06:54:20.917677 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-rs7b8_ce46ec44-fc83-46e0-9f85-159e05452f0f/manager/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.037890 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-88lgf_5904f94f-274a-4b7f-92be-5b8acc5736ab/kube-rbac-proxy/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.165600 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-88lgf_5904f94f-274a-4b7f-92be-5b8acc5736ab/manager/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.212445 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9bc586d8b-5f5p9_e956383f-8ae1-444a-b876-a85b2bcd1bcb/kube-rbac-proxy/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.386364 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7nwnc_4affc2fc-c552-4a1b-b59e-31c7ff2cbb87/kube-rbac-proxy/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.441416 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9bc586d8b-5f5p9_e956383f-8ae1-444a-b876-a85b2bcd1bcb/manager/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.444134 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7nwnc_4affc2fc-c552-4a1b-b59e-31c7ff2cbb87/manager/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.591702 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7wgp_7aa1c48d-537b-4674-a284-1d929fc9cd16/kube-rbac-proxy/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.688941 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7wgp_7aa1c48d-537b-4674-a284-1d929fc9cd16/manager/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.801162 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-8mtps_b87d0440-b095-4540-a0a4-cb71461be4bd/manager/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.832520 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-8mtps_b87d0440-b095-4540-a0a4-cb71461be4bd/kube-rbac-proxy/0.log" Dec 06 06:54:21 crc kubenswrapper[4957]: I1206 06:54:21.869419 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2x6nl_47c9992f-497a-4d57-8c37-c344ad49c199/kube-rbac-proxy/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.000783 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2x6nl_47c9992f-497a-4d57-8c37-c344ad49c199/manager/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.033669 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2cx79_4de58666-a4e1-4cae-99d7-2088192c51db/kube-rbac-proxy/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.120076 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2cx79_4de58666-a4e1-4cae-99d7-2088192c51db/manager/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.260042 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-rrq65_33674637-d0b9-40c9-a482-f4f18c39a80d/kube-rbac-proxy/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.435938 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c2t2c_db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5/manager/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.439903 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c2t2c_db4ead2f-afaf-4f2f-8fc5-5eeb0c4745b5/kube-rbac-proxy/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.487656 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-rrq65_33674637-d0b9-40c9-a482-f4f18c39a80d/manager/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.602664 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8_f9a79638-2a84-4bc8-a3c8-3289f3a1adae/manager/0.log" Dec 06 06:54:22 crc kubenswrapper[4957]: I1206 06:54:22.652067 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4vsfd8_f9a79638-2a84-4bc8-a3c8-3289f3a1adae/kube-rbac-proxy/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.065648 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-854f9c554b-shlk7_87dfe427-cd07-464f-9d50-ca308f553d91/operator/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.079482 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dr5bc_931598ec-e32e-4be7-81cf-1a5993ad1226/registry-server/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.252534 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mc7wf_0a9ebb33-9805-4dec-a879-df7417c8ce7b/kube-rbac-proxy/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.386597 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mc7wf_0a9ebb33-9805-4dec-a879-df7417c8ce7b/manager/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.497502 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-g6zcq_d349e041-2168-4b8a-b6c4-a5840de8c9d8/kube-rbac-proxy/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.520633 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-g6zcq_d349e041-2168-4b8a-b6c4-a5840de8c9d8/manager/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.682328 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-6gccn_31332329-9dd4-4665-b1d0-9662ee3d1b8e/operator/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.773458 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xjbq2_2d7a7ff4-095a-4dff-9c96-25d03888a51e/kube-rbac-proxy/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.893100 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-xjbq2_2d7a7ff4-095a-4dff-9c96-25d03888a51e/manager/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.914057 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-75c7c544f9-bnr4h_b48eed1d-ba6d-41a4-bb23-2d6cc39c4e2e/manager/0.log" Dec 06 06:54:23 crc kubenswrapper[4957]: I1206 06:54:23.932783 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nr7jl_f2ab1e95-3e28-4d99-9340-f1eb24bc6dca/kube-rbac-proxy/0.log" Dec 06 06:54:24 crc kubenswrapper[4957]: I1206 06:54:24.076552 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-nr7jl_f2ab1e95-3e28-4d99-9340-f1eb24bc6dca/manager/0.log" Dec 06 06:54:24 crc kubenswrapper[4957]: I1206 06:54:24.101304 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fvj4l_ea50a02a-23dc-4fcf-bcfd-048b66ce8152/kube-rbac-proxy/0.log" Dec 06 06:54:24 crc kubenswrapper[4957]: I1206 06:54:24.125991 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fvj4l_ea50a02a-23dc-4fcf-bcfd-048b66ce8152/manager/0.log" Dec 06 06:54:24 crc kubenswrapper[4957]: I1206 06:54:24.273587 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k4mm8_ee5ec7b9-749f-4b24-826c-bebb580f7c5b/manager/0.log" Dec 06 06:54:24 crc kubenswrapper[4957]: I1206 06:54:24.283048 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k4mm8_ee5ec7b9-749f-4b24-826c-bebb580f7c5b/kube-rbac-proxy/0.log" Dec 06 06:54:27 crc kubenswrapper[4957]: I1206 06:54:27.663248 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:54:27 crc kubenswrapper[4957]: E1206 06:54:27.663525 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:54:39 crc kubenswrapper[4957]: I1206 06:54:39.662198 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:54:39 crc kubenswrapper[4957]: E1206 06:54:39.662898 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:54:44 crc kubenswrapper[4957]: I1206 06:54:44.798736 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-m7xn8_92ffc396-7acd-4dd1-8bff-61901a7e6159/control-plane-machine-set-operator/0.log" Dec 06 06:54:44 crc kubenswrapper[4957]: I1206 06:54:44.991765 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5ld8_b9ad5cbf-adfe-4012-8968-0d7748db277d/kube-rbac-proxy/0.log" Dec 06 06:54:45 crc kubenswrapper[4957]: I1206 06:54:45.016979 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5ld8_b9ad5cbf-adfe-4012-8968-0d7748db277d/machine-api-operator/0.log" Dec 06 06:54:51 crc kubenswrapper[4957]: I1206 06:54:51.662184 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:54:51 crc kubenswrapper[4957]: E1206 06:54:51.663066 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:54:59 crc kubenswrapper[4957]: I1206 06:54:59.389772 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-59bbt_ff08ac61-0abb-4f45-a647-57aca5172e36/cert-manager-controller/0.log" Dec 06 06:54:59 crc kubenswrapper[4957]: I1206 06:54:59.519573 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-6557k_14f39aaa-5904-4a8d-86ad-ac487456788b/cert-manager-cainjector/0.log" Dec 06 06:54:59 crc kubenswrapper[4957]: I1206 06:54:59.554378 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-b256m_8ebad773-0957-4bb8-b583-a0cff091ff46/cert-manager-webhook/0.log" Dec 06 06:55:04 crc kubenswrapper[4957]: I1206 06:55:04.662618 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:55:04 crc kubenswrapper[4957]: E1206 06:55:04.663343 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:55:12 crc kubenswrapper[4957]: I1206 06:55:12.153243 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-wtc5f_cbf7952f-0cb5-4873-a116-b878ccdbf7e2/nmstate-console-plugin/0.log" Dec 06 06:55:12 crc kubenswrapper[4957]: I1206 06:55:12.329033 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-x8lpv_211db7e0-522a-4635-a279-1d58502c1d1f/nmstate-handler/0.log" Dec 06 06:55:12 crc kubenswrapper[4957]: I1206 06:55:12.355558 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mlsvn_a60b6144-6a1f-4a86-ae5e-84d5771a8c9c/kube-rbac-proxy/0.log" Dec 06 06:55:12 crc kubenswrapper[4957]: I1206 06:55:12.400861 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mlsvn_a60b6144-6a1f-4a86-ae5e-84d5771a8c9c/nmstate-metrics/0.log" Dec 06 06:55:12 crc kubenswrapper[4957]: I1206 06:55:12.568074 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-6r8nd_f85d16c2-33be-4772-a597-fa4393b5846c/nmstate-webhook/0.log" Dec 06 06:55:12 crc kubenswrapper[4957]: I1206 06:55:12.593303 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-2d6j5_44a5fc06-b6e1-4852-a528-f92117991c7a/nmstate-operator/0.log" Dec 06 06:55:18 crc kubenswrapper[4957]: I1206 06:55:18.662966 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:55:18 crc kubenswrapper[4957]: E1206 06:55:18.663771 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:55:27 crc kubenswrapper[4957]: I1206 06:55:27.791340 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzfkk_5ee6dee4-31cd-4d22-905c-a0339c07f1b1/kube-rbac-proxy/0.log" Dec 06 06:55:27 crc kubenswrapper[4957]: I1206 06:55:27.873400 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzfkk_5ee6dee4-31cd-4d22-905c-a0339c07f1b1/controller/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.021788 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.212424 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.234407 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.264805 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.271918 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.477427 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.522241 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.543318 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.544223 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.708434 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-frr-files/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.720221 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-reloader/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.728251 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/cp-metrics/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.750459 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/controller/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.936587 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/frr-metrics/0.log" Dec 06 06:55:28 crc kubenswrapper[4957]: I1206 06:55:28.982820 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/kube-rbac-proxy/0.log" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.000457 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/kube-rbac-proxy-frr/0.log" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.398594 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/reloader/0.log" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.426394 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-sbt5p_5338fe04-633c-405b-8079-1001859b804f/frr-k8s-webhook-server/0.log" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.647145 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9b6b64df4-mn6tm_c4d7ff46-b83b-4551-8aac-8ad1ed6b8b3f/manager/0.log" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.661864 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:55:29 crc kubenswrapper[4957]: E1206 06:55:29.663028 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.820963 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5568f7b8bc-8wsfl_c3dd44fb-8340-44f9-af68-f7e628a59b6c/webhook-server/0.log" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.862911 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2jbdq_fa194652-baa4-424b-bcd0-736ac7295ec4/kube-rbac-proxy/0.log" Dec 06 06:55:29 crc kubenswrapper[4957]: I1206 06:55:29.978987 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mv448_33816fc1-892d-45d9-b755-803b1bf561bf/frr/0.log" Dec 06 06:55:30 crc kubenswrapper[4957]: I1206 06:55:30.256353 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-2jbdq_fa194652-baa4-424b-bcd0-736ac7295ec4/speaker/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.355804 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/util/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.529533 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/pull/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.555382 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/util/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.566448 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/pull/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.737048 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/pull/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.751975 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/util/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.776566 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcvvhn_0c166589-a548-437d-bdeb-b4cefa7fa6bc/extract/0.log" Dec 06 06:55:42 crc kubenswrapper[4957]: I1206 06:55:42.896808 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/util/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.103187 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/pull/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.133123 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/pull/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.193602 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/util/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.387557 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/util/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.393449 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/pull/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.441736 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xc52b_555b7328-82f2-4233-ba2a-de0145c1560b/extract/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.553524 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-utilities/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.772636 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-utilities/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.798894 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-content/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.828359 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-content/0.log" Dec 06 06:55:43 crc kubenswrapper[4957]: I1206 06:55:43.977857 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-utilities/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.068739 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/extract-content/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.407534 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-utilities/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.475298 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-crj2z_5be927bc-2bb0-4676-929c-bfd93bacc0f5/registry-server/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.606950 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-content/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.613549 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-utilities/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.623588 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-content/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.662588 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:55:44 crc kubenswrapper[4957]: E1206 06:55:44.662876 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.812758 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-utilities/0.log" Dec 06 06:55:44 crc kubenswrapper[4957]: I1206 06:55:44.824456 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/extract-content/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.143354 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cbbgh_de45cbbb-4070-4df4-8b0e-c6dbe32a1a33/marketplace-operator/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.195410 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-utilities/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.400370 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-utilities/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.420405 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vsrb4_25410e1a-80c2-4f9f-9c62-2bd939fa21c2/registry-server/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.425301 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-content/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.500934 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-content/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.627506 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-utilities/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.643287 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/extract-content/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.835639 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vqq25_a5c2e60d-d9ec-470a-948b-70bb6184b9f4/registry-server/0.log" Dec 06 06:55:45 crc kubenswrapper[4957]: I1206 06:55:45.916672 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-utilities/0.log" Dec 06 06:55:46 crc kubenswrapper[4957]: I1206 06:55:46.071854 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-utilities/0.log" Dec 06 06:55:46 crc kubenswrapper[4957]: I1206 06:55:46.112050 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-content/0.log" Dec 06 06:55:46 crc kubenswrapper[4957]: I1206 06:55:46.155802 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-content/0.log" Dec 06 06:55:46 crc kubenswrapper[4957]: I1206 06:55:46.330787 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-utilities/0.log" Dec 06 06:55:46 crc kubenswrapper[4957]: I1206 06:55:46.380342 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/extract-content/0.log" Dec 06 06:55:47 crc kubenswrapper[4957]: I1206 06:55:47.014966 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8twxk_4b08140a-aeb1-4f88-a30c-061e3ce9f403/registry-server/0.log" Dec 06 06:55:58 crc kubenswrapper[4957]: I1206 06:55:58.662211 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:55:58 crc kubenswrapper[4957]: E1206 06:55:58.663160 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.191698 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7brtt"] Dec 06 06:56:03 crc kubenswrapper[4957]: E1206 06:56:03.192731 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d3398d-db92-4c25-b214-c3eff574a032" containerName="container-00" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.192749 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d3398d-db92-4c25-b214-c3eff574a032" containerName="container-00" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.193094 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d3398d-db92-4c25-b214-c3eff574a032" containerName="container-00" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.194800 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.203978 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7brtt"] Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.319055 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-utilities\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.319112 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-catalog-content\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.319232 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pr8z\" (UniqueName: \"kubernetes.io/projected/693fa6a5-de50-4a0a-bf5b-f8715770dff6-kube-api-access-8pr8z\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.421447 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pr8z\" (UniqueName: \"kubernetes.io/projected/693fa6a5-de50-4a0a-bf5b-f8715770dff6-kube-api-access-8pr8z\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.421560 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-utilities\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.421589 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-catalog-content\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.422108 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-catalog-content\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.422120 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-utilities\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.595428 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pr8z\" (UniqueName: \"kubernetes.io/projected/693fa6a5-de50-4a0a-bf5b-f8715770dff6-kube-api-access-8pr8z\") pod \"redhat-operators-7brtt\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:03 crc kubenswrapper[4957]: I1206 06:56:03.823116 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:04 crc kubenswrapper[4957]: I1206 06:56:04.369563 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7brtt"] Dec 06 06:56:04 crc kubenswrapper[4957]: I1206 06:56:04.721116 4957 generic.go:334] "Generic (PLEG): container finished" podID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerID="f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32" exitCode=0 Dec 06 06:56:04 crc kubenswrapper[4957]: I1206 06:56:04.721527 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7brtt" event={"ID":"693fa6a5-de50-4a0a-bf5b-f8715770dff6","Type":"ContainerDied","Data":"f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32"} Dec 06 06:56:04 crc kubenswrapper[4957]: I1206 06:56:04.721593 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7brtt" event={"ID":"693fa6a5-de50-4a0a-bf5b-f8715770dff6","Type":"ContainerStarted","Data":"b9c2a976545c540fd2d181a65fde5363e5d9a494bc5c0a58515383283ba2db49"} Dec 06 06:56:06 crc kubenswrapper[4957]: I1206 06:56:06.806332 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7brtt" event={"ID":"693fa6a5-de50-4a0a-bf5b-f8715770dff6","Type":"ContainerStarted","Data":"aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1"} Dec 06 06:56:08 crc kubenswrapper[4957]: I1206 06:56:08.826738 4957 generic.go:334] "Generic (PLEG): container finished" podID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerID="aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1" exitCode=0 Dec 06 06:56:08 crc kubenswrapper[4957]: I1206 06:56:08.826817 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7brtt" event={"ID":"693fa6a5-de50-4a0a-bf5b-f8715770dff6","Type":"ContainerDied","Data":"aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1"} Dec 06 06:56:09 crc kubenswrapper[4957]: I1206 06:56:09.837141 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7brtt" event={"ID":"693fa6a5-de50-4a0a-bf5b-f8715770dff6","Type":"ContainerStarted","Data":"439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21"} Dec 06 06:56:09 crc kubenswrapper[4957]: I1206 06:56:09.861273 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7brtt" podStartSLOduration=2.379668919 podStartE2EDuration="6.861251646s" podCreationTimestamp="2025-12-06 06:56:03 +0000 UTC" firstStartedPulling="2025-12-06 06:56:04.750320245 +0000 UTC m=+4604.400587877" lastFinishedPulling="2025-12-06 06:56:09.231902972 +0000 UTC m=+4608.882170604" observedRunningTime="2025-12-06 06:56:09.856164167 +0000 UTC m=+4609.506431809" watchObservedRunningTime="2025-12-06 06:56:09.861251646 +0000 UTC m=+4609.511519278" Dec 06 06:56:12 crc kubenswrapper[4957]: I1206 06:56:12.661908 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:56:12 crc kubenswrapper[4957]: E1206 06:56:12.662378 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:56:13 crc kubenswrapper[4957]: I1206 06:56:13.823667 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:13 crc kubenswrapper[4957]: I1206 06:56:13.823969 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:14 crc kubenswrapper[4957]: I1206 06:56:14.895354 4957 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7brtt" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="registry-server" probeResult="failure" output=< Dec 06 06:56:14 crc kubenswrapper[4957]: timeout: failed to connect service ":50051" within 1s Dec 06 06:56:14 crc kubenswrapper[4957]: > Dec 06 06:56:23 crc kubenswrapper[4957]: I1206 06:56:23.894908 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:23 crc kubenswrapper[4957]: I1206 06:56:23.956447 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:24 crc kubenswrapper[4957]: I1206 06:56:24.167841 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7brtt"] Dec 06 06:56:24 crc kubenswrapper[4957]: I1206 06:56:24.970584 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7brtt" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="registry-server" containerID="cri-o://439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21" gracePeriod=2 Dec 06 06:56:25 crc kubenswrapper[4957]: E1206 06:56:25.335596 4957 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod693fa6a5_de50_4a0a_bf5b_f8715770dff6.slice/crio-conmon-439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21.scope\": RecentStats: unable to find data in memory cache]" Dec 06 06:56:25 crc kubenswrapper[4957]: I1206 06:56:25.662500 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:56:25 crc kubenswrapper[4957]: E1206 06:56:25.662712 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:56:25 crc kubenswrapper[4957]: I1206 06:56:25.941079 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:25 crc kubenswrapper[4957]: I1206 06:56:25.986716 4957 generic.go:334] "Generic (PLEG): container finished" podID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerID="439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21" exitCode=0 Dec 06 06:56:25 crc kubenswrapper[4957]: I1206 06:56:25.986775 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7brtt" event={"ID":"693fa6a5-de50-4a0a-bf5b-f8715770dff6","Type":"ContainerDied","Data":"439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21"} Dec 06 06:56:25 crc kubenswrapper[4957]: I1206 06:56:25.986816 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7brtt" event={"ID":"693fa6a5-de50-4a0a-bf5b-f8715770dff6","Type":"ContainerDied","Data":"b9c2a976545c540fd2d181a65fde5363e5d9a494bc5c0a58515383283ba2db49"} Dec 06 06:56:25 crc kubenswrapper[4957]: I1206 06:56:25.986867 4957 scope.go:117] "RemoveContainer" containerID="439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21" Dec 06 06:56:25 crc kubenswrapper[4957]: I1206 06:56:25.986988 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7brtt" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.010041 4957 scope.go:117] "RemoveContainer" containerID="aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.047815 4957 scope.go:117] "RemoveContainer" containerID="f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.064448 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-catalog-content\") pod \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.064735 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pr8z\" (UniqueName: \"kubernetes.io/projected/693fa6a5-de50-4a0a-bf5b-f8715770dff6-kube-api-access-8pr8z\") pod \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.064772 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-utilities\") pod \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\" (UID: \"693fa6a5-de50-4a0a-bf5b-f8715770dff6\") " Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.066470 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-utilities" (OuterVolumeSpecName: "utilities") pod "693fa6a5-de50-4a0a-bf5b-f8715770dff6" (UID: "693fa6a5-de50-4a0a-bf5b-f8715770dff6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.072105 4957 scope.go:117] "RemoveContainer" containerID="439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.072575 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693fa6a5-de50-4a0a-bf5b-f8715770dff6-kube-api-access-8pr8z" (OuterVolumeSpecName: "kube-api-access-8pr8z") pod "693fa6a5-de50-4a0a-bf5b-f8715770dff6" (UID: "693fa6a5-de50-4a0a-bf5b-f8715770dff6"). InnerVolumeSpecName "kube-api-access-8pr8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:56:26 crc kubenswrapper[4957]: E1206 06:56:26.073058 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21\": container with ID starting with 439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21 not found: ID does not exist" containerID="439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.073115 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21"} err="failed to get container status \"439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21\": rpc error: code = NotFound desc = could not find container \"439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21\": container with ID starting with 439bd031b8fc9dac40b2b1bac9ead53a7cc824860cd236594964c860be486f21 not found: ID does not exist" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.073146 4957 scope.go:117] "RemoveContainer" containerID="aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1" Dec 06 06:56:26 crc kubenswrapper[4957]: E1206 06:56:26.073540 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1\": container with ID starting with aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1 not found: ID does not exist" containerID="aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.073640 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1"} err="failed to get container status \"aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1\": rpc error: code = NotFound desc = could not find container \"aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1\": container with ID starting with aa852ff9a157d417b2061b37887694072271b6f5f4470a04a8cb41946475a2f1 not found: ID does not exist" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.073670 4957 scope.go:117] "RemoveContainer" containerID="f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32" Dec 06 06:56:26 crc kubenswrapper[4957]: E1206 06:56:26.073999 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32\": container with ID starting with f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32 not found: ID does not exist" containerID="f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.074047 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32"} err="failed to get container status \"f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32\": rpc error: code = NotFound desc = could not find container \"f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32\": container with ID starting with f27df5d08c03c96b60ae4543eed83f562e17c12ae5aa4e65915d1e0ea511cc32 not found: ID does not exist" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.168073 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pr8z\" (UniqueName: \"kubernetes.io/projected/693fa6a5-de50-4a0a-bf5b-f8715770dff6-kube-api-access-8pr8z\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.168097 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.192019 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "693fa6a5-de50-4a0a-bf5b-f8715770dff6" (UID: "693fa6a5-de50-4a0a-bf5b-f8715770dff6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.270892 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/693fa6a5-de50-4a0a-bf5b-f8715770dff6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.331063 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7brtt"] Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.343478 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7brtt"] Dec 06 06:56:26 crc kubenswrapper[4957]: I1206 06:56:26.674433 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" path="/var/lib/kubelet/pods/693fa6a5-de50-4a0a-bf5b-f8715770dff6/volumes" Dec 06 06:56:38 crc kubenswrapper[4957]: I1206 06:56:38.666910 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:56:38 crc kubenswrapper[4957]: E1206 06:56:38.668183 4957 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f2z46_openshift-machine-config-operator(446a0886-35d8-4574-9ccd-6b38f6df37bc)\"" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.587744 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pztqb"] Dec 06 06:56:40 crc kubenswrapper[4957]: E1206 06:56:40.588549 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="extract-content" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.588564 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="extract-content" Dec 06 06:56:40 crc kubenswrapper[4957]: E1206 06:56:40.588600 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="registry-server" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.588612 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="registry-server" Dec 06 06:56:40 crc kubenswrapper[4957]: E1206 06:56:40.588632 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="extract-utilities" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.588640 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="extract-utilities" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.588893 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="693fa6a5-de50-4a0a-bf5b-f8715770dff6" containerName="registry-server" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.590654 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.602652 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pztqb"] Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.651685 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-catalog-content\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.651737 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px7fm\" (UniqueName: \"kubernetes.io/projected/e3ff5e65-51e2-4206-afab-ef088954f9d0-kube-api-access-px7fm\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.651822 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-utilities\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.753701 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-catalog-content\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.753756 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px7fm\" (UniqueName: \"kubernetes.io/projected/e3ff5e65-51e2-4206-afab-ef088954f9d0-kube-api-access-px7fm\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.753793 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-utilities\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.754309 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-catalog-content\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.754338 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-utilities\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.774215 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px7fm\" (UniqueName: \"kubernetes.io/projected/e3ff5e65-51e2-4206-afab-ef088954f9d0-kube-api-access-px7fm\") pod \"community-operators-pztqb\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:40 crc kubenswrapper[4957]: I1206 06:56:40.919126 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:41 crc kubenswrapper[4957]: I1206 06:56:41.492418 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pztqb"] Dec 06 06:56:42 crc kubenswrapper[4957]: I1206 06:56:42.145874 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pztqb" event={"ID":"e3ff5e65-51e2-4206-afab-ef088954f9d0","Type":"ContainerStarted","Data":"3a3f9fa99995d0cdd317bf307c08c19dadc4d41ad5ccb3c99b53886022d8b027"} Dec 06 06:56:43 crc kubenswrapper[4957]: I1206 06:56:43.164327 4957 generic.go:334] "Generic (PLEG): container finished" podID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerID="b59ccf2e12a4d683bb79593fa3a29dc95cadc57750e29e6e6f81812ad1e0a879" exitCode=0 Dec 06 06:56:43 crc kubenswrapper[4957]: I1206 06:56:43.164524 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pztqb" event={"ID":"e3ff5e65-51e2-4206-afab-ef088954f9d0","Type":"ContainerDied","Data":"b59ccf2e12a4d683bb79593fa3a29dc95cadc57750e29e6e6f81812ad1e0a879"} Dec 06 06:56:45 crc kubenswrapper[4957]: I1206 06:56:45.185523 4957 generic.go:334] "Generic (PLEG): container finished" podID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerID="9f6ad3d1f62c1ef96d55e75297b88aa7962c664efe1abd6c20923fc09ca0494f" exitCode=0 Dec 06 06:56:45 crc kubenswrapper[4957]: I1206 06:56:45.185559 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pztqb" event={"ID":"e3ff5e65-51e2-4206-afab-ef088954f9d0","Type":"ContainerDied","Data":"9f6ad3d1f62c1ef96d55e75297b88aa7962c664efe1abd6c20923fc09ca0494f"} Dec 06 06:56:46 crc kubenswrapper[4957]: I1206 06:56:46.199320 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pztqb" event={"ID":"e3ff5e65-51e2-4206-afab-ef088954f9d0","Type":"ContainerStarted","Data":"3a24e731583e64d0d0b1ef9ad42a0030e4bf967689665f2f7647691e61a2617b"} Dec 06 06:56:46 crc kubenswrapper[4957]: I1206 06:56:46.235921 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pztqb" podStartSLOduration=3.812430034 podStartE2EDuration="6.235900358s" podCreationTimestamp="2025-12-06 06:56:40 +0000 UTC" firstStartedPulling="2025-12-06 06:56:43.16845333 +0000 UTC m=+4642.818720962" lastFinishedPulling="2025-12-06 06:56:45.591923654 +0000 UTC m=+4645.242191286" observedRunningTime="2025-12-06 06:56:46.225171576 +0000 UTC m=+4645.875439228" watchObservedRunningTime="2025-12-06 06:56:46.235900358 +0000 UTC m=+4645.886167980" Dec 06 06:56:50 crc kubenswrapper[4957]: I1206 06:56:50.919936 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:50 crc kubenswrapper[4957]: I1206 06:56:50.921131 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:50 crc kubenswrapper[4957]: I1206 06:56:50.977785 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:51 crc kubenswrapper[4957]: I1206 06:56:51.306936 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:51 crc kubenswrapper[4957]: I1206 06:56:51.364590 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pztqb"] Dec 06 06:56:53 crc kubenswrapper[4957]: I1206 06:56:53.265470 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pztqb" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="registry-server" containerID="cri-o://3a24e731583e64d0d0b1ef9ad42a0030e4bf967689665f2f7647691e61a2617b" gracePeriod=2 Dec 06 06:56:53 crc kubenswrapper[4957]: I1206 06:56:53.663942 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.276045 4957 generic.go:334] "Generic (PLEG): container finished" podID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerID="3a24e731583e64d0d0b1ef9ad42a0030e4bf967689665f2f7647691e61a2617b" exitCode=0 Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.276130 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pztqb" event={"ID":"e3ff5e65-51e2-4206-afab-ef088954f9d0","Type":"ContainerDied","Data":"3a24e731583e64d0d0b1ef9ad42a0030e4bf967689665f2f7647691e61a2617b"} Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.277108 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pztqb" event={"ID":"e3ff5e65-51e2-4206-afab-ef088954f9d0","Type":"ContainerDied","Data":"3a3f9fa99995d0cdd317bf307c08c19dadc4d41ad5ccb3c99b53886022d8b027"} Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.277175 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a3f9fa99995d0cdd317bf307c08c19dadc4d41ad5ccb3c99b53886022d8b027" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.493327 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.589338 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px7fm\" (UniqueName: \"kubernetes.io/projected/e3ff5e65-51e2-4206-afab-ef088954f9d0-kube-api-access-px7fm\") pod \"e3ff5e65-51e2-4206-afab-ef088954f9d0\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.589706 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-utilities\") pod \"e3ff5e65-51e2-4206-afab-ef088954f9d0\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.589754 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-catalog-content\") pod \"e3ff5e65-51e2-4206-afab-ef088954f9d0\" (UID: \"e3ff5e65-51e2-4206-afab-ef088954f9d0\") " Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.590906 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-utilities" (OuterVolumeSpecName: "utilities") pod "e3ff5e65-51e2-4206-afab-ef088954f9d0" (UID: "e3ff5e65-51e2-4206-afab-ef088954f9d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.595910 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ff5e65-51e2-4206-afab-ef088954f9d0-kube-api-access-px7fm" (OuterVolumeSpecName: "kube-api-access-px7fm") pod "e3ff5e65-51e2-4206-afab-ef088954f9d0" (UID: "e3ff5e65-51e2-4206-afab-ef088954f9d0"). InnerVolumeSpecName "kube-api-access-px7fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.636116 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3ff5e65-51e2-4206-afab-ef088954f9d0" (UID: "e3ff5e65-51e2-4206-afab-ef088954f9d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.695293 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px7fm\" (UniqueName: \"kubernetes.io/projected/e3ff5e65-51e2-4206-afab-ef088954f9d0-kube-api-access-px7fm\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.695340 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:54 crc kubenswrapper[4957]: I1206 06:56:54.695353 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3ff5e65-51e2-4206-afab-ef088954f9d0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:56:55 crc kubenswrapper[4957]: I1206 06:56:55.289238 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pztqb" Dec 06 06:56:55 crc kubenswrapper[4957]: I1206 06:56:55.290415 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"c1b8ea828664e9209aa17e32031ed658d1720dc4e2e438130e3ca1e9d9b7f0ae"} Dec 06 06:56:55 crc kubenswrapper[4957]: I1206 06:56:55.371900 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pztqb"] Dec 06 06:56:55 crc kubenswrapper[4957]: I1206 06:56:55.383211 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pztqb"] Dec 06 06:56:56 crc kubenswrapper[4957]: I1206 06:56:56.679015 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" path="/var/lib/kubelet/pods/e3ff5e65-51e2-4206-afab-ef088954f9d0/volumes" Dec 06 06:57:33 crc kubenswrapper[4957]: I1206 06:57:33.657869 4957 generic.go:334] "Generic (PLEG): container finished" podID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerID="e100d4da4629ad227b749c1fee0c2c8df1a72baf374b059a17134b6569ebed57" exitCode=0 Dec 06 06:57:33 crc kubenswrapper[4957]: I1206 06:57:33.657989 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" event={"ID":"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a","Type":"ContainerDied","Data":"e100d4da4629ad227b749c1fee0c2c8df1a72baf374b059a17134b6569ebed57"} Dec 06 06:57:33 crc kubenswrapper[4957]: I1206 06:57:33.659551 4957 scope.go:117] "RemoveContainer" containerID="e100d4da4629ad227b749c1fee0c2c8df1a72baf374b059a17134b6569ebed57" Dec 06 06:57:34 crc kubenswrapper[4957]: I1206 06:57:34.040142 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9qkxf_must-gather-8sgxr_bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a/gather/0.log" Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.437891 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9qkxf/must-gather-8sgxr"] Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.438966 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerName="copy" containerID="cri-o://b8044f1232b7628997571c1e8f10ce4e153db9779f26dcb3dcaedb1ef12b8bbc" gracePeriod=2 Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.447134 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9qkxf/must-gather-8sgxr"] Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.781275 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9qkxf_must-gather-8sgxr_bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a/copy/0.log" Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.782634 4957 generic.go:334] "Generic (PLEG): container finished" podID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerID="b8044f1232b7628997571c1e8f10ce4e153db9779f26dcb3dcaedb1ef12b8bbc" exitCode=143 Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.870929 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9qkxf_must-gather-8sgxr_bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a/copy/0.log" Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.871468 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.926698 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-must-gather-output\") pod \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.927272 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnsq7\" (UniqueName: \"kubernetes.io/projected/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-kube-api-access-lnsq7\") pod \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\" (UID: \"bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a\") " Dec 06 06:57:43 crc kubenswrapper[4957]: I1206 06:57:43.933473 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-kube-api-access-lnsq7" (OuterVolumeSpecName: "kube-api-access-lnsq7") pod "bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" (UID: "bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a"). InnerVolumeSpecName "kube-api-access-lnsq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.029690 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnsq7\" (UniqueName: \"kubernetes.io/projected/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-kube-api-access-lnsq7\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.096531 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" (UID: "bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.131763 4957 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.672523 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" path="/var/lib/kubelet/pods/bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a/volumes" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.791698 4957 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9qkxf_must-gather-8sgxr_bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a/copy/0.log" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.792679 4957 scope.go:117] "RemoveContainer" containerID="b8044f1232b7628997571c1e8f10ce4e153db9779f26dcb3dcaedb1ef12b8bbc" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.792732 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9qkxf/must-gather-8sgxr" Dec 06 06:57:44 crc kubenswrapper[4957]: I1206 06:57:44.817544 4957 scope.go:117] "RemoveContainer" containerID="e100d4da4629ad227b749c1fee0c2c8df1a72baf374b059a17134b6569ebed57" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.888347 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4f56z"] Dec 06 06:58:22 crc kubenswrapper[4957]: E1206 06:58:22.889637 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="extract-content" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.889657 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="extract-content" Dec 06 06:58:22 crc kubenswrapper[4957]: E1206 06:58:22.889682 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="extract-utilities" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.889691 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="extract-utilities" Dec 06 06:58:22 crc kubenswrapper[4957]: E1206 06:58:22.889710 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerName="copy" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.889718 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerName="copy" Dec 06 06:58:22 crc kubenswrapper[4957]: E1206 06:58:22.889732 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="registry-server" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.889739 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="registry-server" Dec 06 06:58:22 crc kubenswrapper[4957]: E1206 06:58:22.889766 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerName="gather" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.889773 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerName="gather" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.890036 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerName="copy" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.890064 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf3a9a3f-2a91-4d6e-a3be-b610fbfe510a" containerName="gather" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.890122 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ff5e65-51e2-4206-afab-ef088954f9d0" containerName="registry-server" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.891933 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:22 crc kubenswrapper[4957]: I1206 06:58:22.904430 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4f56z"] Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.029690 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-utilities\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.029739 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsm8g\" (UniqueName: \"kubernetes.io/projected/ae575291-f688-4efa-b555-cdbd0d877dcc-kube-api-access-dsm8g\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.029811 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-catalog-content\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.132784 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-utilities\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.132858 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsm8g\" (UniqueName: \"kubernetes.io/projected/ae575291-f688-4efa-b555-cdbd0d877dcc-kube-api-access-dsm8g\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.132929 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-catalog-content\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.133415 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-catalog-content\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.133415 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-utilities\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.172576 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsm8g\" (UniqueName: \"kubernetes.io/projected/ae575291-f688-4efa-b555-cdbd0d877dcc-kube-api-access-dsm8g\") pod \"redhat-marketplace-4f56z\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.214732 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:23 crc kubenswrapper[4957]: I1206 06:58:23.696381 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4f56z"] Dec 06 06:58:24 crc kubenswrapper[4957]: I1206 06:58:24.549872 4957 generic.go:334] "Generic (PLEG): container finished" podID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerID="4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef" exitCode=0 Dec 06 06:58:24 crc kubenswrapper[4957]: I1206 06:58:24.549973 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4f56z" event={"ID":"ae575291-f688-4efa-b555-cdbd0d877dcc","Type":"ContainerDied","Data":"4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef"} Dec 06 06:58:24 crc kubenswrapper[4957]: I1206 06:58:24.550205 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4f56z" event={"ID":"ae575291-f688-4efa-b555-cdbd0d877dcc","Type":"ContainerStarted","Data":"6cf74cd88e363f319062686759b0fee04a74c88ea8a87ef560c59b9d70bd7c98"} Dec 06 06:58:24 crc kubenswrapper[4957]: I1206 06:58:24.552719 4957 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 06:58:25 crc kubenswrapper[4957]: I1206 06:58:25.563134 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4f56z" event={"ID":"ae575291-f688-4efa-b555-cdbd0d877dcc","Type":"ContainerStarted","Data":"d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185"} Dec 06 06:58:26 crc kubenswrapper[4957]: I1206 06:58:26.579400 4957 generic.go:334] "Generic (PLEG): container finished" podID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerID="d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185" exitCode=0 Dec 06 06:58:26 crc kubenswrapper[4957]: I1206 06:58:26.579442 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4f56z" event={"ID":"ae575291-f688-4efa-b555-cdbd0d877dcc","Type":"ContainerDied","Data":"d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185"} Dec 06 06:58:27 crc kubenswrapper[4957]: I1206 06:58:27.589894 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4f56z" event={"ID":"ae575291-f688-4efa-b555-cdbd0d877dcc","Type":"ContainerStarted","Data":"1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5"} Dec 06 06:58:27 crc kubenswrapper[4957]: I1206 06:58:27.621601 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4f56z" podStartSLOduration=3.16648324 podStartE2EDuration="5.621575237s" podCreationTimestamp="2025-12-06 06:58:22 +0000 UTC" firstStartedPulling="2025-12-06 06:58:24.552280488 +0000 UTC m=+4744.202548120" lastFinishedPulling="2025-12-06 06:58:27.007372485 +0000 UTC m=+4746.657640117" observedRunningTime="2025-12-06 06:58:27.610868305 +0000 UTC m=+4747.261135947" watchObservedRunningTime="2025-12-06 06:58:27.621575237 +0000 UTC m=+4747.271842889" Dec 06 06:58:33 crc kubenswrapper[4957]: I1206 06:58:33.215567 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:33 crc kubenswrapper[4957]: I1206 06:58:33.216298 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:33 crc kubenswrapper[4957]: I1206 06:58:33.282725 4957 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:33 crc kubenswrapper[4957]: I1206 06:58:33.715993 4957 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:33 crc kubenswrapper[4957]: I1206 06:58:33.764899 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4f56z"] Dec 06 06:58:35 crc kubenswrapper[4957]: I1206 06:58:35.673974 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4f56z" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="registry-server" containerID="cri-o://1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5" gracePeriod=2 Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.158557 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.300963 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-catalog-content\") pod \"ae575291-f688-4efa-b555-cdbd0d877dcc\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.301140 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsm8g\" (UniqueName: \"kubernetes.io/projected/ae575291-f688-4efa-b555-cdbd0d877dcc-kube-api-access-dsm8g\") pod \"ae575291-f688-4efa-b555-cdbd0d877dcc\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.301209 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-utilities\") pod \"ae575291-f688-4efa-b555-cdbd0d877dcc\" (UID: \"ae575291-f688-4efa-b555-cdbd0d877dcc\") " Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.302461 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-utilities" (OuterVolumeSpecName: "utilities") pod "ae575291-f688-4efa-b555-cdbd0d877dcc" (UID: "ae575291-f688-4efa-b555-cdbd0d877dcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.309430 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae575291-f688-4efa-b555-cdbd0d877dcc-kube-api-access-dsm8g" (OuterVolumeSpecName: "kube-api-access-dsm8g") pod "ae575291-f688-4efa-b555-cdbd0d877dcc" (UID: "ae575291-f688-4efa-b555-cdbd0d877dcc"). InnerVolumeSpecName "kube-api-access-dsm8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.323086 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae575291-f688-4efa-b555-cdbd0d877dcc" (UID: "ae575291-f688-4efa-b555-cdbd0d877dcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.403923 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsm8g\" (UniqueName: \"kubernetes.io/projected/ae575291-f688-4efa-b555-cdbd0d877dcc-kube-api-access-dsm8g\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.403977 4957 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.403994 4957 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae575291-f688-4efa-b555-cdbd0d877dcc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.692408 4957 generic.go:334] "Generic (PLEG): container finished" podID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerID="1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5" exitCode=0 Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.692474 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4f56z" event={"ID":"ae575291-f688-4efa-b555-cdbd0d877dcc","Type":"ContainerDied","Data":"1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5"} Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.692502 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4f56z" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.692521 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4f56z" event={"ID":"ae575291-f688-4efa-b555-cdbd0d877dcc","Type":"ContainerDied","Data":"6cf74cd88e363f319062686759b0fee04a74c88ea8a87ef560c59b9d70bd7c98"} Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.692545 4957 scope.go:117] "RemoveContainer" containerID="1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.725072 4957 scope.go:117] "RemoveContainer" containerID="d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.755152 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4f56z"] Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.764797 4957 scope.go:117] "RemoveContainer" containerID="4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.765535 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4f56z"] Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.816277 4957 scope.go:117] "RemoveContainer" containerID="1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5" Dec 06 06:58:36 crc kubenswrapper[4957]: E1206 06:58:36.816848 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5\": container with ID starting with 1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5 not found: ID does not exist" containerID="1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.816899 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5"} err="failed to get container status \"1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5\": rpc error: code = NotFound desc = could not find container \"1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5\": container with ID starting with 1eb022d73713eddde806217c6f3d384bdfdeee6654665525031b31761152f2d5 not found: ID does not exist" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.816928 4957 scope.go:117] "RemoveContainer" containerID="d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185" Dec 06 06:58:36 crc kubenswrapper[4957]: E1206 06:58:36.817419 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185\": container with ID starting with d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185 not found: ID does not exist" containerID="d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.817441 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185"} err="failed to get container status \"d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185\": rpc error: code = NotFound desc = could not find container \"d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185\": container with ID starting with d8c53f50cd4dba193561835b139cb61f1a6d356c5ed26d819c446c7967de6185 not found: ID does not exist" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.817455 4957 scope.go:117] "RemoveContainer" containerID="4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef" Dec 06 06:58:36 crc kubenswrapper[4957]: E1206 06:58:36.817755 4957 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef\": container with ID starting with 4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef not found: ID does not exist" containerID="4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef" Dec 06 06:58:36 crc kubenswrapper[4957]: I1206 06:58:36.817799 4957 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef"} err="failed to get container status \"4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef\": rpc error: code = NotFound desc = could not find container \"4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef\": container with ID starting with 4d41ce0aa8b973d394b66a662f4d956a2a77b3fc12ffcec7f12dae8f92555fef not found: ID does not exist" Dec 06 06:58:38 crc kubenswrapper[4957]: I1206 06:58:38.682908 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" path="/var/lib/kubelet/pods/ae575291-f688-4efa-b555-cdbd0d877dcc/volumes" Dec 06 06:59:10 crc kubenswrapper[4957]: I1206 06:59:10.181552 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:59:10 crc kubenswrapper[4957]: I1206 06:59:10.182294 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 06:59:40 crc kubenswrapper[4957]: I1206 06:59:40.181276 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 06:59:40 crc kubenswrapper[4957]: I1206 06:59:40.182011 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.175520 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb"] Dec 06 07:00:00 crc kubenswrapper[4957]: E1206 07:00:00.177229 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="registry-server" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.177253 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="registry-server" Dec 06 07:00:00 crc kubenswrapper[4957]: E1206 07:00:00.177309 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="extract-utilities" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.177324 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="extract-utilities" Dec 06 07:00:00 crc kubenswrapper[4957]: E1206 07:00:00.177356 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="extract-content" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.177367 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="extract-content" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.177598 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae575291-f688-4efa-b555-cdbd0d877dcc" containerName="registry-server" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.178657 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.181077 4957 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.183218 4957 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.188149 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb"] Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.375381 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltwnm\" (UniqueName: \"kubernetes.io/projected/b28f0de4-64c4-41dc-b832-853ca86c1a29-kube-api-access-ltwnm\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.375530 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b28f0de4-64c4-41dc-b832-853ca86c1a29-secret-volume\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.377532 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b28f0de4-64c4-41dc-b832-853ca86c1a29-config-volume\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.479685 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltwnm\" (UniqueName: \"kubernetes.io/projected/b28f0de4-64c4-41dc-b832-853ca86c1a29-kube-api-access-ltwnm\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.479773 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b28f0de4-64c4-41dc-b832-853ca86c1a29-secret-volume\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.479948 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b28f0de4-64c4-41dc-b832-853ca86c1a29-config-volume\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.481033 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b28f0de4-64c4-41dc-b832-853ca86c1a29-config-volume\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.489578 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b28f0de4-64c4-41dc-b832-853ca86c1a29-secret-volume\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.527867 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltwnm\" (UniqueName: \"kubernetes.io/projected/b28f0de4-64c4-41dc-b832-853ca86c1a29-kube-api-access-ltwnm\") pod \"collect-profiles-29416740-jc6mb\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:00 crc kubenswrapper[4957]: I1206 07:00:00.808554 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:01 crc kubenswrapper[4957]: I1206 07:00:01.294130 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb"] Dec 06 07:00:01 crc kubenswrapper[4957]: I1206 07:00:01.534614 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" event={"ID":"b28f0de4-64c4-41dc-b832-853ca86c1a29","Type":"ContainerStarted","Data":"634347f1d35322081812631f577ea767fa1266d1f33d800f24d717842dabf292"} Dec 06 07:00:02 crc kubenswrapper[4957]: I1206 07:00:02.550588 4957 generic.go:334] "Generic (PLEG): container finished" podID="b28f0de4-64c4-41dc-b832-853ca86c1a29" containerID="0cfef37c1d906dddfecf9d465f77d50a88d8ec1ced2f5035c26192c901651bbc" exitCode=0 Dec 06 07:00:02 crc kubenswrapper[4957]: I1206 07:00:02.550660 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" event={"ID":"b28f0de4-64c4-41dc-b832-853ca86c1a29","Type":"ContainerDied","Data":"0cfef37c1d906dddfecf9d465f77d50a88d8ec1ced2f5035c26192c901651bbc"} Dec 06 07:00:02 crc kubenswrapper[4957]: I1206 07:00:02.574796 4957 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5789568449-4t258" podUID="475d23a9-0dda-4ba5-8af9-108574997a4f" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 06 07:00:03 crc kubenswrapper[4957]: I1206 07:00:03.920400 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.063100 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b28f0de4-64c4-41dc-b832-853ca86c1a29-secret-volume\") pod \"b28f0de4-64c4-41dc-b832-853ca86c1a29\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.063567 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltwnm\" (UniqueName: \"kubernetes.io/projected/b28f0de4-64c4-41dc-b832-853ca86c1a29-kube-api-access-ltwnm\") pod \"b28f0de4-64c4-41dc-b832-853ca86c1a29\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.063617 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b28f0de4-64c4-41dc-b832-853ca86c1a29-config-volume\") pod \"b28f0de4-64c4-41dc-b832-853ca86c1a29\" (UID: \"b28f0de4-64c4-41dc-b832-853ca86c1a29\") " Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.064486 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b28f0de4-64c4-41dc-b832-853ca86c1a29-config-volume" (OuterVolumeSpecName: "config-volume") pod "b28f0de4-64c4-41dc-b832-853ca86c1a29" (UID: "b28f0de4-64c4-41dc-b832-853ca86c1a29"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.066746 4957 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b28f0de4-64c4-41dc-b832-853ca86c1a29-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.069899 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b28f0de4-64c4-41dc-b832-853ca86c1a29-kube-api-access-ltwnm" (OuterVolumeSpecName: "kube-api-access-ltwnm") pod "b28f0de4-64c4-41dc-b832-853ca86c1a29" (UID: "b28f0de4-64c4-41dc-b832-853ca86c1a29"). InnerVolumeSpecName "kube-api-access-ltwnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.070340 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b28f0de4-64c4-41dc-b832-853ca86c1a29-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b28f0de4-64c4-41dc-b832-853ca86c1a29" (UID: "b28f0de4-64c4-41dc-b832-853ca86c1a29"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.168704 4957 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b28f0de4-64c4-41dc-b832-853ca86c1a29-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.168993 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltwnm\" (UniqueName: \"kubernetes.io/projected/b28f0de4-64c4-41dc-b832-853ca86c1a29-kube-api-access-ltwnm\") on node \"crc\" DevicePath \"\"" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.568740 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" event={"ID":"b28f0de4-64c4-41dc-b832-853ca86c1a29","Type":"ContainerDied","Data":"634347f1d35322081812631f577ea767fa1266d1f33d800f24d717842dabf292"} Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.568786 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="634347f1d35322081812631f577ea767fa1266d1f33d800f24d717842dabf292" Dec 06 07:00:04 crc kubenswrapper[4957]: I1206 07:00:04.568789 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416740-jc6mb" Dec 06 07:00:05 crc kubenswrapper[4957]: I1206 07:00:05.002986 4957 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz"] Dec 06 07:00:05 crc kubenswrapper[4957]: I1206 07:00:05.012523 4957 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416695-cffcz"] Dec 06 07:00:06 crc kubenswrapper[4957]: I1206 07:00:06.673281 4957 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bde94ee6-96a0-4433-9b7f-73a34c6085c8" path="/var/lib/kubelet/pods/bde94ee6-96a0-4433-9b7f-73a34c6085c8/volumes" Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.180816 4957 patch_prober.go:28] interesting pod/machine-config-daemon-f2z46 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.181218 4957 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.181277 4957 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.182163 4957 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1b8ea828664e9209aa17e32031ed658d1720dc4e2e438130e3ca1e9d9b7f0ae"} pod="openshift-machine-config-operator/machine-config-daemon-f2z46" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.182231 4957 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" podUID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerName="machine-config-daemon" containerID="cri-o://c1b8ea828664e9209aa17e32031ed658d1720dc4e2e438130e3ca1e9d9b7f0ae" gracePeriod=600 Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.630014 4957 generic.go:334] "Generic (PLEG): container finished" podID="446a0886-35d8-4574-9ccd-6b38f6df37bc" containerID="c1b8ea828664e9209aa17e32031ed658d1720dc4e2e438130e3ca1e9d9b7f0ae" exitCode=0 Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.630332 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerDied","Data":"c1b8ea828664e9209aa17e32031ed658d1720dc4e2e438130e3ca1e9d9b7f0ae"} Dec 06 07:00:10 crc kubenswrapper[4957]: I1206 07:00:10.630469 4957 scope.go:117] "RemoveContainer" containerID="a1a574c4a2f8fb092a1c3cdb002f1acda549eb9e8a0a8e206d951d7f58cd355a" Dec 06 07:00:11 crc kubenswrapper[4957]: I1206 07:00:11.661349 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f2z46" event={"ID":"446a0886-35d8-4574-9ccd-6b38f6df37bc","Type":"ContainerStarted","Data":"78236c516949f3c2ae9ee6f2cd71abcfd9b896fa9eeb71b301084192c720373c"} Dec 06 07:00:46 crc kubenswrapper[4957]: I1206 07:00:46.567587 4957 scope.go:117] "RemoveContainer" containerID="a761b3c2de714eeb28749d7301343ae955b5c06956633df234b0bf7e07a258e1" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.163878 4957 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29416741-d7brg"] Dec 06 07:01:00 crc kubenswrapper[4957]: E1206 07:01:00.165418 4957 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b28f0de4-64c4-41dc-b832-853ca86c1a29" containerName="collect-profiles" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.165437 4957 state_mem.go:107] "Deleted CPUSet assignment" podUID="b28f0de4-64c4-41dc-b832-853ca86c1a29" containerName="collect-profiles" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.165714 4957 memory_manager.go:354] "RemoveStaleState removing state" podUID="b28f0de4-64c4-41dc-b832-853ca86c1a29" containerName="collect-profiles" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.166811 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.200868 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416741-d7brg"] Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.234626 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-combined-ca-bundle\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.235099 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n4zr\" (UniqueName: \"kubernetes.io/projected/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-kube-api-access-5n4zr\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.235159 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-config-data\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.235230 4957 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-fernet-keys\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.336210 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-fernet-keys\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.336271 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-combined-ca-bundle\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.336394 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n4zr\" (UniqueName: \"kubernetes.io/projected/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-kube-api-access-5n4zr\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.336455 4957 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-config-data\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.343904 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-combined-ca-bundle\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.344066 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-config-data\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.346794 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-fernet-keys\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.361068 4957 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n4zr\" (UniqueName: \"kubernetes.io/projected/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-kube-api-access-5n4zr\") pod \"keystone-cron-29416741-d7brg\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:00 crc kubenswrapper[4957]: I1206 07:01:00.518086 4957 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:01 crc kubenswrapper[4957]: I1206 07:01:01.000036 4957 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416741-d7brg"] Dec 06 07:01:01 crc kubenswrapper[4957]: W1206 07:01:01.000918 4957 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a7d8b1_f5f6_46dc_bd7d_063ed80d436d.slice/crio-29e039a809ab1b0ad877b565031cdea65d65b470defa1ad544d793e92ef581df WatchSource:0}: Error finding container 29e039a809ab1b0ad877b565031cdea65d65b470defa1ad544d793e92ef581df: Status 404 returned error can't find the container with id 29e039a809ab1b0ad877b565031cdea65d65b470defa1ad544d793e92ef581df Dec 06 07:01:01 crc kubenswrapper[4957]: I1206 07:01:01.214819 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-d7brg" event={"ID":"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d","Type":"ContainerStarted","Data":"fb3f8d93f2d5611487cd864ab21e8b42d29e6a1a66b30486f7c70fc87c7fc7ab"} Dec 06 07:01:01 crc kubenswrapper[4957]: I1206 07:01:01.216326 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-d7brg" event={"ID":"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d","Type":"ContainerStarted","Data":"29e039a809ab1b0ad877b565031cdea65d65b470defa1ad544d793e92ef581df"} Dec 06 07:01:01 crc kubenswrapper[4957]: I1206 07:01:01.240912 4957 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29416741-d7brg" podStartSLOduration=1.240892397 podStartE2EDuration="1.240892397s" podCreationTimestamp="2025-12-06 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 07:01:01.23550218 +0000 UTC m=+4900.885769822" watchObservedRunningTime="2025-12-06 07:01:01.240892397 +0000 UTC m=+4900.891160029" Dec 06 07:01:04 crc kubenswrapper[4957]: I1206 07:01:04.242698 4957 generic.go:334] "Generic (PLEG): container finished" podID="10a7d8b1-f5f6-46dc-bd7d-063ed80d436d" containerID="fb3f8d93f2d5611487cd864ab21e8b42d29e6a1a66b30486f7c70fc87c7fc7ab" exitCode=0 Dec 06 07:01:04 crc kubenswrapper[4957]: I1206 07:01:04.242849 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-d7brg" event={"ID":"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d","Type":"ContainerDied","Data":"fb3f8d93f2d5611487cd864ab21e8b42d29e6a1a66b30486f7c70fc87c7fc7ab"} Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.721012 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-d7brg" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.748588 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-combined-ca-bundle\") pod \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.748741 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-fernet-keys\") pod \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.748818 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n4zr\" (UniqueName: \"kubernetes.io/projected/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-kube-api-access-5n4zr\") pod \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.748927 4957 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-config-data\") pod \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\" (UID: \"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d\") " Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.755233 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-kube-api-access-5n4zr" (OuterVolumeSpecName: "kube-api-access-5n4zr") pod "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d" (UID: "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d"). InnerVolumeSpecName "kube-api-access-5n4zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.756059 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d" (UID: "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.817996 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d" (UID: "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.842165 4957 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-config-data" (OuterVolumeSpecName: "config-data") pod "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d" (UID: "10a7d8b1-f5f6-46dc-bd7d-063ed80d436d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.851027 4957 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.851071 4957 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n4zr\" (UniqueName: \"kubernetes.io/projected/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-kube-api-access-5n4zr\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.851086 4957 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:05 crc kubenswrapper[4957]: I1206 07:01:05.851098 4957 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a7d8b1-f5f6-46dc-bd7d-063ed80d436d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 07:01:06 crc kubenswrapper[4957]: I1206 07:01:06.269880 4957 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416741-d7brg" event={"ID":"10a7d8b1-f5f6-46dc-bd7d-063ed80d436d","Type":"ContainerDied","Data":"29e039a809ab1b0ad877b565031cdea65d65b470defa1ad544d793e92ef581df"} Dec 06 07:01:06 crc kubenswrapper[4957]: I1206 07:01:06.270134 4957 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29e039a809ab1b0ad877b565031cdea65d65b470defa1ad544d793e92ef581df" Dec 06 07:01:06 crc kubenswrapper[4957]: I1206 07:01:06.270193 4957 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416741-d7brg" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114752321024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114752322017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114740232016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114740232015454 5ustar corecore